id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
23,634
def make_api_version(version_data): from readthedocs.builds.models import Version for key in ['resource_uri', 'absolute_url', 'downloads']: if (key in version_data): del version_data[key] project_data = version_data['project'] project = make_api_project(project_data) version_data['project'] = project ver = Version(**version_data) ver.save = _new_save return ver
[ "def", "make_api_version", "(", "version_data", ")", ":", "from", "readthedocs", ".", "builds", ".", "models", "import", "Version", "for", "key", "in", "[", "'resource_uri'", ",", "'absolute_url'", ",", "'downloads'", "]", ":", "if", "(", "key", "in", "version_data", ")", ":", "del", "version_data", "[", "key", "]", "project_data", "=", "version_data", "[", "'project'", "]", "project", "=", "make_api_project", "(", "project_data", ")", "version_data", "[", "'project'", "]", "=", "project", "ver", "=", "Version", "(", "**", "version_data", ")", "ver", ".", "save", "=", "_new_save", "return", "ver" ]
make mock version instance from api return .
train
false
23,635
def assert_version_header_matches_request(api_microversion_header_name, api_microversion, response_header): api_microversion_header_name = api_microversion_header_name.lower() if ((api_microversion_header_name not in response_header) or (api_microversion != response_header[api_microversion_header_name])): msg = ("Microversion header '%s' with value '%s' does not match in response - %s. " % (api_microversion_header_name, api_microversion, response_header)) raise exceptions.InvalidHTTPResponseHeader(msg)
[ "def", "assert_version_header_matches_request", "(", "api_microversion_header_name", ",", "api_microversion", ",", "response_header", ")", ":", "api_microversion_header_name", "=", "api_microversion_header_name", ".", "lower", "(", ")", "if", "(", "(", "api_microversion_header_name", "not", "in", "response_header", ")", "or", "(", "api_microversion", "!=", "response_header", "[", "api_microversion_header_name", "]", ")", ")", ":", "msg", "=", "(", "\"Microversion header '%s' with value '%s' does not match in response - %s. \"", "%", "(", "api_microversion_header_name", ",", "api_microversion", ",", "response_header", ")", ")", "raise", "exceptions", ".", "InvalidHTTPResponseHeader", "(", "msg", ")" ]
checks api microversion in response header verify whether microversion is present in response header and with specified api_microversion value .
train
false
23,636
def _win32_find_exe(exe): candidates = [exe] if (u'.' not in exe): extensions = getenv(u'PATHEXT', u'').split(os.pathsep) candidates.extend([(exe + ext) for ext in extensions if ext.startswith(u'.')]) for candidate in candidates: if exists(candidate): return candidate if (not os.path.dirname(exe)): for path in getenv(u'PATH').split(os.pathsep): if path: for candidate in candidates: full_path = os.path.join(path, candidate) if exists(full_path): return full_path return exe
[ "def", "_win32_find_exe", "(", "exe", ")", ":", "candidates", "=", "[", "exe", "]", "if", "(", "u'.'", "not", "in", "exe", ")", ":", "extensions", "=", "getenv", "(", "u'PATHEXT'", ",", "u''", ")", ".", "split", "(", "os", ".", "pathsep", ")", "candidates", ".", "extend", "(", "[", "(", "exe", "+", "ext", ")", "for", "ext", "in", "extensions", "if", "ext", ".", "startswith", "(", "u'.'", ")", "]", ")", "for", "candidate", "in", "candidates", ":", "if", "exists", "(", "candidate", ")", ":", "return", "candidate", "if", "(", "not", "os", ".", "path", ".", "dirname", "(", "exe", ")", ")", ":", "for", "path", "in", "getenv", "(", "u'PATH'", ")", ".", "split", "(", "os", ".", "pathsep", ")", ":", "if", "path", ":", "for", "candidate", "in", "candidates", ":", "full_path", "=", "os", ".", "path", ".", "join", "(", "path", ",", "candidate", ")", "if", "exists", "(", "full_path", ")", ":", "return", "full_path", "return", "exe" ]
find the actual file for a windows executable .
train
false
23,638
@csrf_exempt def update_subproject(request, project, subproject): if (not appsettings.ENABLE_HOOKS): return HttpResponseNotAllowed([]) obj = get_subproject(request, project, subproject, True) if (not obj.project.enable_hooks): return HttpResponseNotAllowed([]) perform_update(obj) return hook_response()
[ "@", "csrf_exempt", "def", "update_subproject", "(", "request", ",", "project", ",", "subproject", ")", ":", "if", "(", "not", "appsettings", ".", "ENABLE_HOOKS", ")", ":", "return", "HttpResponseNotAllowed", "(", "[", "]", ")", "obj", "=", "get_subproject", "(", "request", ",", "project", ",", "subproject", ",", "True", ")", "if", "(", "not", "obj", ".", "project", ".", "enable_hooks", ")", ":", "return", "HttpResponseNotAllowed", "(", "[", "]", ")", "perform_update", "(", "obj", ")", "return", "hook_response", "(", ")" ]
api hook for updating git repos .
train
false
23,639
def _json_play_players(play, data): players = OrderedDict() for (playerid, statcats) in data.iteritems(): if (playerid == '0'): continue for info in statcats: if (info['statId'] not in nflgame.statmap.idmap): continue if (playerid not in players): home = play.drive.game.is_home(info['clubcode']) if home: team_name = play.drive.game.home else: team_name = play.drive.game.away stats = nflgame.player.PlayPlayerStats(playerid, info['playerName'], home, team_name) players[playerid] = stats statvals = nflgame.statmap.values(info['statId'], info['yards']) players[playerid]._add_stats(statvals) return players
[ "def", "_json_play_players", "(", "play", ",", "data", ")", ":", "players", "=", "OrderedDict", "(", ")", "for", "(", "playerid", ",", "statcats", ")", "in", "data", ".", "iteritems", "(", ")", ":", "if", "(", "playerid", "==", "'0'", ")", ":", "continue", "for", "info", "in", "statcats", ":", "if", "(", "info", "[", "'statId'", "]", "not", "in", "nflgame", ".", "statmap", ".", "idmap", ")", ":", "continue", "if", "(", "playerid", "not", "in", "players", ")", ":", "home", "=", "play", ".", "drive", ".", "game", ".", "is_home", "(", "info", "[", "'clubcode'", "]", ")", "if", "home", ":", "team_name", "=", "play", ".", "drive", ".", "game", ".", "home", "else", ":", "team_name", "=", "play", ".", "drive", ".", "game", ".", "away", "stats", "=", "nflgame", ".", "player", ".", "PlayPlayerStats", "(", "playerid", ",", "info", "[", "'playerName'", "]", ",", "home", ",", "team_name", ")", "players", "[", "playerid", "]", "=", "stats", "statvals", "=", "nflgame", ".", "statmap", ".", "values", "(", "info", "[", "'statId'", "]", ",", "info", "[", "'yards'", "]", ")", "players", "[", "playerid", "]", ".", "_add_stats", "(", "statvals", ")", "return", "players" ]
takes a single json play entry and converts it to an ordereddict of player statistics .
train
false
23,642
def create_list_of(class_, objects): return [class_.from_metadata(obj) for obj in objects]
[ "def", "create_list_of", "(", "class_", ",", "objects", ")", ":", "return", "[", "class_", ".", "from_metadata", "(", "obj", ")", "for", "obj", "in", "objects", "]" ]
return a list of model objects of class class_ from list of object metadata objects .
train
false
23,643
def test_ast_bad_global(): cant_compile(u'(global)') cant_compile(u'(global (foo))')
[ "def", "test_ast_bad_global", "(", ")", ":", "cant_compile", "(", "u'(global)'", ")", "cant_compile", "(", "u'(global (foo))'", ")" ]
make sure ast cant compile invalid global .
train
false
23,644
def parse_clademodelc(branch_type_no, line_floats, site_classes): if ((not site_classes) or (len(line_floats) == 0)): return for n in range(len(line_floats)): if (site_classes[n].get('branch types') is None): site_classes[n]['branch types'] = {} site_classes[n]['branch types'][branch_type_no] = line_floats[n] return site_classes
[ "def", "parse_clademodelc", "(", "branch_type_no", ",", "line_floats", ",", "site_classes", ")", ":", "if", "(", "(", "not", "site_classes", ")", "or", "(", "len", "(", "line_floats", ")", "==", "0", ")", ")", ":", "return", "for", "n", "in", "range", "(", "len", "(", "line_floats", ")", ")", ":", "if", "(", "site_classes", "[", "n", "]", ".", "get", "(", "'branch types'", ")", "is", "None", ")", ":", "site_classes", "[", "n", "]", "[", "'branch types'", "]", "=", "{", "}", "site_classes", "[", "n", "]", "[", "'branch types'", "]", "[", "branch_type_no", "]", "=", "line_floats", "[", "n", "]", "return", "site_classes" ]
parse results specific to the clade model c .
train
false
23,646
def safe_creation(context, create_fn, delete_fn, create_bindings, transaction=True): cm = (context.session.begin(subtransactions=True) if transaction else _noop_context_manager()) with cm: obj = create_fn() try: value = create_bindings(obj['id']) except Exception: with excutils.save_and_reraise_exception(): try: delete_fn(obj['id']) except Exception as e: LOG.error(_LE('Cannot clean up created object %(obj)s. Exception: %(exc)s'), {'obj': obj['id'], 'exc': e}) return (obj, value)
[ "def", "safe_creation", "(", "context", ",", "create_fn", ",", "delete_fn", ",", "create_bindings", ",", "transaction", "=", "True", ")", ":", "cm", "=", "(", "context", ".", "session", ".", "begin", "(", "subtransactions", "=", "True", ")", "if", "transaction", "else", "_noop_context_manager", "(", ")", ")", "with", "cm", ":", "obj", "=", "create_fn", "(", ")", "try", ":", "value", "=", "create_bindings", "(", "obj", "[", "'id'", "]", ")", "except", "Exception", ":", "with", "excutils", ".", "save_and_reraise_exception", "(", ")", ":", "try", ":", "delete_fn", "(", "obj", "[", "'id'", "]", ")", "except", "Exception", "as", "e", ":", "LOG", ".", "error", "(", "_LE", "(", "'Cannot clean up created object %(obj)s. Exception: %(exc)s'", ")", ",", "{", "'obj'", ":", "obj", "[", "'id'", "]", ",", "'exc'", ":", "e", "}", ")", "return", "(", "obj", ",", "value", ")" ]
this function wraps logic of object creation in safe atomic way .
train
false
23,647
@require_POST @login_required @permitted def endorse_comment(request, course_id, comment_id): course_key = CourseKey.from_string(course_id) comment = cc.Comment.find(comment_id) user = request.user comment.endorsed = (request.POST.get('endorsed', 'false').lower() == 'true') comment.endorsement_user_id = user.id comment.save() comment_endorsed.send(sender=None, user=user, post=comment) return JsonResponse(prepare_content(comment.to_dict(), course_key))
[ "@", "require_POST", "@", "login_required", "@", "permitted", "def", "endorse_comment", "(", "request", ",", "course_id", ",", "comment_id", ")", ":", "course_key", "=", "CourseKey", ".", "from_string", "(", "course_id", ")", "comment", "=", "cc", ".", "Comment", ".", "find", "(", "comment_id", ")", "user", "=", "request", ".", "user", "comment", ".", "endorsed", "=", "(", "request", ".", "POST", ".", "get", "(", "'endorsed'", ",", "'false'", ")", ".", "lower", "(", ")", "==", "'true'", ")", "comment", ".", "endorsement_user_id", "=", "user", ".", "id", "comment", ".", "save", "(", ")", "comment_endorsed", ".", "send", "(", "sender", "=", "None", ",", "user", "=", "user", ",", "post", "=", "comment", ")", "return", "JsonResponse", "(", "prepare_content", "(", "comment", ".", "to_dict", "(", ")", ",", "course_key", ")", ")" ]
given a course_id and comment_id .
train
false
23,650
def org_organisation_address(row): if hasattr(row, 'org_organisation'): row = row.org_organisation try: organisation_id = row.id except: return current.messages['NONE'] db = current.db s3db = current.s3db otable = s3db.org_office gtable = s3db.gis_location query = (((otable.deleted != True) & (otable.organisation_id == organisation_id)) & (otable.location_id == gtable.id)) row = db(query).select(gtable.addr_street, limitby=(0, 1)).first() if row: row.addr_street else: return current.messages['NONE']
[ "def", "org_organisation_address", "(", "row", ")", ":", "if", "hasattr", "(", "row", ",", "'org_organisation'", ")", ":", "row", "=", "row", ".", "org_organisation", "try", ":", "organisation_id", "=", "row", ".", "id", "except", ":", "return", "current", ".", "messages", "[", "'NONE'", "]", "db", "=", "current", ".", "db", "s3db", "=", "current", ".", "s3db", "otable", "=", "s3db", ".", "org_office", "gtable", "=", "s3db", ".", "gis_location", "query", "=", "(", "(", "(", "otable", ".", "deleted", "!=", "True", ")", "&", "(", "otable", ".", "organisation_id", "==", "organisation_id", ")", ")", "&", "(", "otable", ".", "location_id", "==", "gtable", ".", "id", ")", ")", "row", "=", "db", "(", "query", ")", ".", "select", "(", "gtable", ".", "addr_street", ",", "limitby", "=", "(", "0", ",", "1", ")", ")", ".", "first", "(", ")", "if", "row", ":", "row", ".", "addr_street", "else", ":", "return", "current", ".", "messages", "[", "'NONE'", "]" ]
the address of the first office .
train
false
23,651
def get_icon_themes(): themes = [] icon_themes_env = core.getenv(u'GIT_COLA_ICON_THEME') if icon_themes_env: themes.extend([x for x in icon_themes_env.split(u':') if x]) icon_themes_cfg = gitcfg.current().get_all(u'cola.icontheme') if icon_themes_cfg: themes.extend(icon_themes_cfg) if (not themes): themes.append(u'light') return themes
[ "def", "get_icon_themes", "(", ")", ":", "themes", "=", "[", "]", "icon_themes_env", "=", "core", ".", "getenv", "(", "u'GIT_COLA_ICON_THEME'", ")", "if", "icon_themes_env", ":", "themes", ".", "extend", "(", "[", "x", "for", "x", "in", "icon_themes_env", ".", "split", "(", "u':'", ")", "if", "x", "]", ")", "icon_themes_cfg", "=", "gitcfg", ".", "current", "(", ")", ".", "get_all", "(", "u'cola.icontheme'", ")", "if", "icon_themes_cfg", ":", "themes", ".", "extend", "(", "icon_themes_cfg", ")", "if", "(", "not", "themes", ")", ":", "themes", ".", "append", "(", "u'light'", ")", "return", "themes" ]
return the default icon theme names .
train
false
23,653
def getUnicode(value, encoding=None, noneToNull=False): if (noneToNull and (value is None)): return NULL if isinstance(value, unicode): return value elif isinstance(value, basestring): while True: try: return unicode(value, (encoding or (kb.get('pageEncoding') if kb.get('originalPage') else None) or UNICODE_ENCODING)) except UnicodeDecodeError as ex: try: return unicode(value, UNICODE_ENCODING) except: value = ((value[:ex.start] + ''.join(((INVALID_UNICODE_CHAR_FORMAT % ord(_)) for _ in value[ex.start:ex.end]))) + value[ex.end:]) elif isListLike(value): value = list((getUnicode(_, encoding, noneToNull) for _ in value)) return value else: try: return unicode(value) except UnicodeDecodeError: return unicode(str(value), errors='ignore')
[ "def", "getUnicode", "(", "value", ",", "encoding", "=", "None", ",", "noneToNull", "=", "False", ")", ":", "if", "(", "noneToNull", "and", "(", "value", "is", "None", ")", ")", ":", "return", "NULL", "if", "isinstance", "(", "value", ",", "unicode", ")", ":", "return", "value", "elif", "isinstance", "(", "value", ",", "basestring", ")", ":", "while", "True", ":", "try", ":", "return", "unicode", "(", "value", ",", "(", "encoding", "or", "(", "kb", ".", "get", "(", "'pageEncoding'", ")", "if", "kb", ".", "get", "(", "'originalPage'", ")", "else", "None", ")", "or", "UNICODE_ENCODING", ")", ")", "except", "UnicodeDecodeError", "as", "ex", ":", "try", ":", "return", "unicode", "(", "value", ",", "UNICODE_ENCODING", ")", "except", ":", "value", "=", "(", "(", "value", "[", ":", "ex", ".", "start", "]", "+", "''", ".", "join", "(", "(", "(", "INVALID_UNICODE_CHAR_FORMAT", "%", "ord", "(", "_", ")", ")", "for", "_", "in", "value", "[", "ex", ".", "start", ":", "ex", ".", "end", "]", ")", ")", ")", "+", "value", "[", "ex", ".", "end", ":", "]", ")", "elif", "isListLike", "(", "value", ")", ":", "value", "=", "list", "(", "(", "getUnicode", "(", "_", ",", "encoding", ",", "noneToNull", ")", "for", "_", "in", "value", ")", ")", "return", "value", "else", ":", "try", ":", "return", "unicode", "(", "value", ")", "except", "UnicodeDecodeError", ":", "return", "unicode", "(", "str", "(", "value", ")", ",", "errors", "=", "'ignore'", ")" ]
return the unicode representation of the supplied value: .
train
false
23,654
def get_custom_data_point(): length = random.randint(0, 10) print 'reporting timeseries value {}'.format(str(length)) return length
[ "def", "get_custom_data_point", "(", ")", ":", "length", "=", "random", ".", "randint", "(", "0", ",", "10", ")", "print", "'reporting timeseries value {}'", ".", "format", "(", "str", "(", "length", ")", ")", "return", "length" ]
dummy method to return a mock measurement for demonstration purposes .
train
false
23,656
def add_accept_handler(sock, callback, io_loop=None): if (io_loop is None): io_loop = IOLoop.current() def accept_handler(fd, events): while True: try: (connection, address) = sock.accept() except socket.error as e: if (e.args[0] in (errno.EWOULDBLOCK, errno.EAGAIN)): return if (e.args[0] == errno.ECONNABORTED): continue raise callback(connection, address) io_loop.add_handler(sock.fileno(), accept_handler, IOLoop.READ)
[ "def", "add_accept_handler", "(", "sock", ",", "callback", ",", "io_loop", "=", "None", ")", ":", "if", "(", "io_loop", "is", "None", ")", ":", "io_loop", "=", "IOLoop", ".", "current", "(", ")", "def", "accept_handler", "(", "fd", ",", "events", ")", ":", "while", "True", ":", "try", ":", "(", "connection", ",", "address", ")", "=", "sock", ".", "accept", "(", ")", "except", "socket", ".", "error", "as", "e", ":", "if", "(", "e", ".", "args", "[", "0", "]", "in", "(", "errno", ".", "EWOULDBLOCK", ",", "errno", ".", "EAGAIN", ")", ")", ":", "return", "if", "(", "e", ".", "args", "[", "0", "]", "==", "errno", ".", "ECONNABORTED", ")", ":", "continue", "raise", "callback", "(", "connection", ",", "address", ")", "io_loop", ".", "add_handler", "(", "sock", ".", "fileno", "(", ")", ",", "accept_handler", ",", "IOLoop", ".", "READ", ")" ]
adds an .
train
false
23,658
def hosted_zone_absent(name, domain_name=None, region=None, key=None, keyid=None, profile=None): domain_name = (domain_name if domain_name else name) ret = {'name': name, 'result': True, 'comment': '', 'changes': {}} deets = __salt__['boto_route53.describe_hosted_zones'](domain_name=domain_name, region=region, key=key, keyid=keyid, profile=profile) if (not deets): ret['comment'] = 'Hosted Zone {0} already absent'.format(domain_name) log.info(ret['comment']) return ret if __opts__['test']: ret['comment'] = 'Route53 Hosted Zone {0} set to be deleted.'.format(domain_name) ret['result'] = None return ret if __salt__['boto_route53.delete_zone'](zone=domain_name, region=region, key=key, keyid=keyid, profile=profile): ret['comment'] = 'Route53 Hosted Zone {0} deleted'.format(domain_name) log.info(ret['comment']) ret['changes']['old'] = deets ret['changes']['new'] = None return ret
[ "def", "hosted_zone_absent", "(", "name", ",", "domain_name", "=", "None", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "domain_name", "=", "(", "domain_name", "if", "domain_name", "else", "name", ")", "ret", "=", "{", "'name'", ":", "name", ",", "'result'", ":", "True", ",", "'comment'", ":", "''", ",", "'changes'", ":", "{", "}", "}", "deets", "=", "__salt__", "[", "'boto_route53.describe_hosted_zones'", "]", "(", "domain_name", "=", "domain_name", ",", "region", "=", "region", ",", "key", "=", "key", ",", "keyid", "=", "keyid", ",", "profile", "=", "profile", ")", "if", "(", "not", "deets", ")", ":", "ret", "[", "'comment'", "]", "=", "'Hosted Zone {0} already absent'", ".", "format", "(", "domain_name", ")", "log", ".", "info", "(", "ret", "[", "'comment'", "]", ")", "return", "ret", "if", "__opts__", "[", "'test'", "]", ":", "ret", "[", "'comment'", "]", "=", "'Route53 Hosted Zone {0} set to be deleted.'", ".", "format", "(", "domain_name", ")", "ret", "[", "'result'", "]", "=", "None", "return", "ret", "if", "__salt__", "[", "'boto_route53.delete_zone'", "]", "(", "zone", "=", "domain_name", ",", "region", "=", "region", ",", "key", "=", "key", ",", "keyid", "=", "keyid", ",", "profile", "=", "profile", ")", ":", "ret", "[", "'comment'", "]", "=", "'Route53 Hosted Zone {0} deleted'", ".", "format", "(", "domain_name", ")", "log", ".", "info", "(", "ret", "[", "'comment'", "]", ")", "ret", "[", "'changes'", "]", "[", "'old'", "]", "=", "deets", "ret", "[", "'changes'", "]", "[", "'new'", "]", "=", "None", "return", "ret" ]
ensure the route53 hostes zone described is absent name the name of the state definition .
train
true
23,659
def classImplements(cls, *interfaces): spec = implementedBy(cls) spec.declared += tuple(_normalizeargs(interfaces)) bases = [] seen = {} for b in spec.declared: if (b not in seen): seen[b] = 1 bases.append(b) if (spec.inherit is not None): for c in spec.inherit.__bases__: b = implementedBy(c) if (b not in seen): seen[b] = 1 bases.append(b) spec.__bases__ = tuple(bases)
[ "def", "classImplements", "(", "cls", ",", "*", "interfaces", ")", ":", "spec", "=", "implementedBy", "(", "cls", ")", "spec", ".", "declared", "+=", "tuple", "(", "_normalizeargs", "(", "interfaces", ")", ")", "bases", "=", "[", "]", "seen", "=", "{", "}", "for", "b", "in", "spec", ".", "declared", ":", "if", "(", "b", "not", "in", "seen", ")", ":", "seen", "[", "b", "]", "=", "1", "bases", ".", "append", "(", "b", ")", "if", "(", "spec", ".", "inherit", "is", "not", "None", ")", ":", "for", "c", "in", "spec", ".", "inherit", ".", "__bases__", ":", "b", "=", "implementedBy", "(", "c", ")", "if", "(", "b", "not", "in", "seen", ")", ":", "seen", "[", "b", "]", "=", "1", "bases", ".", "append", "(", "b", ")", "spec", ".", "__bases__", "=", "tuple", "(", "bases", ")" ]
declare additional interfaces implemented for instances of a class the arguments after the class are one or more interfaces or interface specifications .
train
false
23,660
def _api_queue(name, output, kwargs): value = kwargs.get('value', '') return _api_queue_table.get(name, (_api_queue_default, 2))[0](output, value, kwargs)
[ "def", "_api_queue", "(", "name", ",", "output", ",", "kwargs", ")", ":", "value", "=", "kwargs", ".", "get", "(", "'value'", ",", "''", ")", "return", "_api_queue_table", ".", "get", "(", "name", ",", "(", "_api_queue_default", ",", "2", ")", ")", "[", "0", "]", "(", "output", ",", "value", ",", "kwargs", ")" ]
api: dispatcher for mode=queue .
train
false
23,661
def vq(obs, code_book, check_finite=True): obs = _asarray_validated(obs, check_finite=check_finite) code_book = _asarray_validated(code_book, check_finite=check_finite) ct = np.common_type(obs, code_book) c_obs = obs.astype(ct, copy=False) if (code_book.dtype != ct): c_code_book = code_book.astype(ct) else: c_code_book = code_book if (ct in (np.float32, np.float64)): results = _vq.vq(c_obs, c_code_book) else: results = py_vq(obs, code_book) return results
[ "def", "vq", "(", "obs", ",", "code_book", ",", "check_finite", "=", "True", ")", ":", "obs", "=", "_asarray_validated", "(", "obs", ",", "check_finite", "=", "check_finite", ")", "code_book", "=", "_asarray_validated", "(", "code_book", ",", "check_finite", "=", "check_finite", ")", "ct", "=", "np", ".", "common_type", "(", "obs", ",", "code_book", ")", "c_obs", "=", "obs", ".", "astype", "(", "ct", ",", "copy", "=", "False", ")", "if", "(", "code_book", ".", "dtype", "!=", "ct", ")", ":", "c_code_book", "=", "code_book", ".", "astype", "(", "ct", ")", "else", ":", "c_code_book", "=", "code_book", "if", "(", "ct", "in", "(", "np", ".", "float32", ",", "np", ".", "float64", ")", ")", ":", "results", "=", "_vq", ".", "vq", "(", "c_obs", ",", "c_code_book", ")", "else", ":", "results", "=", "py_vq", "(", "obs", ",", "code_book", ")", "return", "results" ]
assign codes from a code book to observations .
train
false
23,666
def _api_restart(name, output, kwargs): sabnzbd.trigger_restart() return report(output)
[ "def", "_api_restart", "(", "name", ",", "output", ",", "kwargs", ")", ":", "sabnzbd", ".", "trigger_restart", "(", ")", "return", "report", "(", "output", ")" ]
api: accepts output .
train
false
23,667
def story(z, image_loc, k=100, bw=50, lyric=False): (rawim, im) = load_image(image_loc) feats = compute_features(z['net'], im).flatten() feats /= norm(feats) feats = embedding.encode_images(z['vse'], feats[None, :]) scores = numpy.dot(feats, z['cvec'].T).flatten() sorted_args = numpy.argsort(scores)[::(-1)] sentences = [z['cap'][a] for a in sorted_args[:k]] print 'NEAREST-CAPTIONS: ' for s in sentences[:5]: print s print '' svecs = skipthoughts.encode(z['stv'], sentences, verbose=False) shift = ((svecs.mean(0) - z['bneg']) + z['bpos']) passage = decoder.run_sampler(z['dec'], shift, beam_width=bw) print 'OUTPUT: ' if lyric: for line in passage.split(','): if (line[0] != ' '): print line else: print line[1:] else: print passage
[ "def", "story", "(", "z", ",", "image_loc", ",", "k", "=", "100", ",", "bw", "=", "50", ",", "lyric", "=", "False", ")", ":", "(", "rawim", ",", "im", ")", "=", "load_image", "(", "image_loc", ")", "feats", "=", "compute_features", "(", "z", "[", "'net'", "]", ",", "im", ")", ".", "flatten", "(", ")", "feats", "/=", "norm", "(", "feats", ")", "feats", "=", "embedding", ".", "encode_images", "(", "z", "[", "'vse'", "]", ",", "feats", "[", "None", ",", ":", "]", ")", "scores", "=", "numpy", ".", "dot", "(", "feats", ",", "z", "[", "'cvec'", "]", ".", "T", ")", ".", "flatten", "(", ")", "sorted_args", "=", "numpy", ".", "argsort", "(", "scores", ")", "[", ":", ":", "(", "-", "1", ")", "]", "sentences", "=", "[", "z", "[", "'cap'", "]", "[", "a", "]", "for", "a", "in", "sorted_args", "[", ":", "k", "]", "]", "print", "'NEAREST-CAPTIONS: '", "for", "s", "in", "sentences", "[", ":", "5", "]", ":", "print", "s", "print", "''", "svecs", "=", "skipthoughts", ".", "encode", "(", "z", "[", "'stv'", "]", ",", "sentences", ",", "verbose", "=", "False", ")", "shift", "=", "(", "(", "svecs", ".", "mean", "(", "0", ")", "-", "z", "[", "'bneg'", "]", ")", "+", "z", "[", "'bpos'", "]", ")", "passage", "=", "decoder", ".", "run_sampler", "(", "z", "[", "'dec'", "]", ",", "shift", ",", "beam_width", "=", "bw", ")", "print", "'OUTPUT: '", "if", "lyric", ":", "for", "line", "in", "passage", ".", "split", "(", "','", ")", ":", "if", "(", "line", "[", "0", "]", "!=", "' '", ")", ":", "print", "line", "else", ":", "print", "line", "[", "1", ":", "]", "else", ":", "print", "passage" ]
generate a story for an image at location image_loc .
train
false
23,668
def check_string_length(value, name, min_length=0, max_length=None, allow_all_spaces=True): try: strutils.check_string_length(value, name=name, min_length=min_length, max_length=max_length) except (ValueError, TypeError) as exc: raise exception.InvalidInput(reason=exc) if ((not allow_all_spaces) and value.isspace()): msg = _('%(name)s cannot be all spaces.') raise exception.InvalidInput(reason=msg)
[ "def", "check_string_length", "(", "value", ",", "name", ",", "min_length", "=", "0", ",", "max_length", "=", "None", ",", "allow_all_spaces", "=", "True", ")", ":", "try", ":", "strutils", ".", "check_string_length", "(", "value", ",", "name", "=", "name", ",", "min_length", "=", "min_length", ",", "max_length", "=", "max_length", ")", "except", "(", "ValueError", ",", "TypeError", ")", "as", "exc", ":", "raise", "exception", ".", "InvalidInput", "(", "reason", "=", "exc", ")", "if", "(", "(", "not", "allow_all_spaces", ")", "and", "value", ".", "isspace", "(", ")", ")", ":", "msg", "=", "_", "(", "'%(name)s cannot be all spaces.'", ")", "raise", "exception", ".", "InvalidInput", "(", "reason", "=", "msg", ")" ]
check the length of specified string .
train
false
23,669
def getNewRepository(): return ExportRepository()
[ "def", "getNewRepository", "(", ")", ":", "return", "ExportRepository", "(", ")" ]
get new repository .
train
false
23,670
def is_x64_architecture(): if ('64' in platform.machine()): return True else: return False
[ "def", "is_x64_architecture", "(", ")", ":", "if", "(", "'64'", "in", "platform", ".", "machine", "(", ")", ")", ":", "return", "True", "else", ":", "return", "False" ]
return true if the architecture is x64 .
train
false
23,671
@gen.engine def DetermineStartDate(client, job, callback): start_date = options.options.start_date if options.options.smart_scan: last_run = (yield gen.Task(job.FindLastSuccess, with_payload_key='stats.last_day')) if (last_run is None): logging.info('No previous successful scan found, rerun with --start_date') callback(None) return last_run_start = last_run['start_time'] if ((last_run_start + (options.options.hours_between_runs * constants.SECONDS_PER_HOUR)) > time.time()): logging.info(('Last successful run started at %s, less than %d hours ago; skipping.' % (time.asctime(time.localtime(last_run_start)), options.options.hours_between_runs))) callback(None) return last_day = last_run['stats.last_day'] start_time = (util.ISO8601ToUTCTimestamp(last_day) + constants.SECONDS_PER_DAY) start_date = util.TimestampUTCToISO8601(start_time) logging.info(('Last successful run (%s) scanned up to %s, setting start date to %s' % (time.asctime(time.localtime(last_run_start)), last_day, start_date))) callback(start_date)
[ "@", "gen", ".", "engine", "def", "DetermineStartDate", "(", "client", ",", "job", ",", "callback", ")", ":", "start_date", "=", "options", ".", "options", ".", "start_date", "if", "options", ".", "options", ".", "smart_scan", ":", "last_run", "=", "(", "yield", "gen", ".", "Task", "(", "job", ".", "FindLastSuccess", ",", "with_payload_key", "=", "'stats.last_day'", ")", ")", "if", "(", "last_run", "is", "None", ")", ":", "logging", ".", "info", "(", "'No previous successful scan found, rerun with --start_date'", ")", "callback", "(", "None", ")", "return", "last_run_start", "=", "last_run", "[", "'start_time'", "]", "if", "(", "(", "last_run_start", "+", "(", "options", ".", "options", ".", "hours_between_runs", "*", "constants", ".", "SECONDS_PER_HOUR", ")", ")", ">", "time", ".", "time", "(", ")", ")", ":", "logging", ".", "info", "(", "(", "'Last successful run started at %s, less than %d hours ago; skipping.'", "%", "(", "time", ".", "asctime", "(", "time", ".", "localtime", "(", "last_run_start", ")", ")", ",", "options", ".", "options", ".", "hours_between_runs", ")", ")", ")", "callback", "(", "None", ")", "return", "last_day", "=", "last_run", "[", "'stats.last_day'", "]", "start_time", "=", "(", "util", ".", "ISO8601ToUTCTimestamp", "(", "last_day", ")", "+", "constants", ".", "SECONDS_PER_DAY", ")", "start_date", "=", "util", ".", "TimestampUTCToISO8601", "(", "start_time", ")", "logging", ".", "info", "(", "(", "'Last successful run (%s) scanned up to %s, setting start date to %s'", "%", "(", "time", ".", "asctime", "(", "time", ".", "localtime", "(", "last_run_start", ")", ")", ",", "last_day", ",", "start_date", ")", ")", ")", "callback", "(", "start_date", ")" ]
if smart_scan is true .
train
false
23,672
@pytest.fixture() def make_app(test_params): apps = [] syspath = sys.path[:] def make(*args, **kwargs): (status, warning) = (StringIO(), StringIO()) kwargs.setdefault('status', status) kwargs.setdefault('warning', warning) app_ = util.SphinxTestApp(*args, **kwargs) apps.append(app_) if test_params['shared_result']: app_ = SphinxTestAppWrapperForSkipBuilding(app_) return app_ (yield make) sys.path[:] = syspath for app_ in apps: app_.cleanup()
[ "@", "pytest", ".", "fixture", "(", ")", "def", "make_app", "(", "test_params", ")", ":", "apps", "=", "[", "]", "syspath", "=", "sys", ".", "path", "[", ":", "]", "def", "make", "(", "*", "args", ",", "**", "kwargs", ")", ":", "(", "status", ",", "warning", ")", "=", "(", "StringIO", "(", ")", ",", "StringIO", "(", ")", ")", "kwargs", ".", "setdefault", "(", "'status'", ",", "status", ")", "kwargs", ".", "setdefault", "(", "'warning'", ",", "warning", ")", "app_", "=", "util", ".", "SphinxTestApp", "(", "*", "args", ",", "**", "kwargs", ")", "apps", ".", "append", "(", "app_", ")", "if", "test_params", "[", "'shared_result'", "]", ":", "app_", "=", "SphinxTestAppWrapperForSkipBuilding", "(", "app_", ")", "return", "app_", "(", "yield", "make", ")", "sys", ".", "path", "[", ":", "]", "=", "syspath", "for", "app_", "in", "apps", ":", "app_", ".", "cleanup", "(", ")" ]
factory function that creates a new coolmagicapplication object .
train
false
23,673
def _count(a, axis=None): if hasattr(a, 'count'): num = a.count(axis=axis) if (isinstance(num, np.ndarray) and (num.ndim == 0)): num = int(num) elif (axis is None): num = a.size else: num = a.shape[axis] return num
[ "def", "_count", "(", "a", ",", "axis", "=", "None", ")", ":", "if", "hasattr", "(", "a", ",", "'count'", ")", ":", "num", "=", "a", ".", "count", "(", "axis", "=", "axis", ")", "if", "(", "isinstance", "(", "num", ",", "np", ".", "ndarray", ")", "and", "(", "num", ".", "ndim", "==", "0", ")", ")", ":", "num", "=", "int", "(", "num", ")", "elif", "(", "axis", "is", "None", ")", ":", "num", "=", "a", ".", "size", "else", ":", "num", "=", "a", ".", "shape", "[", "axis", "]", "return", "num" ]
count the number of non-masked elements of an array .
train
false
23,674
@utils.arg('host', metavar='<hostname>', help=_('Name of host.')) @utils.arg('--status', metavar='<enable|disable>', default=None, dest='status', help=_('Either enable or disable a host.')) @utils.arg('--maintenance', metavar='<enable|disable>', default=None, dest='maintenance', help=_('Either put or resume host to/from maintenance.')) def do_host_update(cs, args): updates = {} columns = ['HOST'] if args.status: updates['status'] = args.status columns.append('status') if args.maintenance: updates['maintenance_mode'] = args.maintenance columns.append('maintenance_mode') result = cs.hosts.update(args.host, updates) utils.print_list([result], columns)
[ "@", "utils", ".", "arg", "(", "'host'", ",", "metavar", "=", "'<hostname>'", ",", "help", "=", "_", "(", "'Name of host.'", ")", ")", "@", "utils", ".", "arg", "(", "'--status'", ",", "metavar", "=", "'<enable|disable>'", ",", "default", "=", "None", ",", "dest", "=", "'status'", ",", "help", "=", "_", "(", "'Either enable or disable a host.'", ")", ")", "@", "utils", ".", "arg", "(", "'--maintenance'", ",", "metavar", "=", "'<enable|disable>'", ",", "default", "=", "None", ",", "dest", "=", "'maintenance'", ",", "help", "=", "_", "(", "'Either put or resume host to/from maintenance.'", ")", ")", "def", "do_host_update", "(", "cs", ",", "args", ")", ":", "updates", "=", "{", "}", "columns", "=", "[", "'HOST'", "]", "if", "args", ".", "status", ":", "updates", "[", "'status'", "]", "=", "args", ".", "status", "columns", ".", "append", "(", "'status'", ")", "if", "args", ".", "maintenance", ":", "updates", "[", "'maintenance_mode'", "]", "=", "args", ".", "maintenance", "columns", ".", "append", "(", "'maintenance_mode'", ")", "result", "=", "cs", ".", "hosts", ".", "update", "(", "args", ".", "host", ",", "updates", ")", "utils", ".", "print_list", "(", "[", "result", "]", ",", "columns", ")" ]
update host settings .
train
false
23,677
def overload(func): return _overload_dummy
[ "def", "overload", "(", "func", ")", ":", "return", "_overload_dummy" ]
a decorator marking the decorated function as typing and implementing *func* in nopython mode .
train
false
23,678
def _set_users(users): return __salt__['users.set_users'](users, commit=False)
[ "def", "_set_users", "(", "users", ")", ":", "return", "__salt__", "[", "'users.set_users'", "]", "(", "users", ",", "commit", "=", "False", ")" ]
calls users .
train
false
23,679
def create_settings_file(init=True): settings_path = os.path.join(GAMEDIR, 'server', 'conf', 'settings.py') if (not init): if os.path.exists(settings_path): inp = raw_input('server/conf/settings.py already exists. Do you want to reset it? y/[N]> ') if (not (inp.lower() == 'y')): print('Aborted.') sys.exit() else: print('Reset the settings file.') default_settings_path = os.path.join(EVENNIA_TEMPLATE, 'server', 'conf', 'settings.py') shutil.copy(default_settings_path, settings_path) with open(settings_path, 'r') as f: settings_string = f.read() setting_dict = {'settings_default': os.path.join(EVENNIA_LIB, 'settings_default.py'), 'servername': ('"%s"' % GAMEDIR.rsplit(os.path.sep, 1)[1].capitalize()), 'secret_key': ("'%s'" % create_secret_key())} settings_string = settings_string.format(**setting_dict) with open(settings_path, 'w') as f: f.write(settings_string)
[ "def", "create_settings_file", "(", "init", "=", "True", ")", ":", "settings_path", "=", "os", ".", "path", ".", "join", "(", "GAMEDIR", ",", "'server'", ",", "'conf'", ",", "'settings.py'", ")", "if", "(", "not", "init", ")", ":", "if", "os", ".", "path", ".", "exists", "(", "settings_path", ")", ":", "inp", "=", "raw_input", "(", "'server/conf/settings.py already exists. Do you want to reset it? y/[N]> '", ")", "if", "(", "not", "(", "inp", ".", "lower", "(", ")", "==", "'y'", ")", ")", ":", "print", "(", "'Aborted.'", ")", "sys", ".", "exit", "(", ")", "else", ":", "print", "(", "'Reset the settings file.'", ")", "default_settings_path", "=", "os", ".", "path", ".", "join", "(", "EVENNIA_TEMPLATE", ",", "'server'", ",", "'conf'", ",", "'settings.py'", ")", "shutil", ".", "copy", "(", "default_settings_path", ",", "settings_path", ")", "with", "open", "(", "settings_path", ",", "'r'", ")", "as", "f", ":", "settings_string", "=", "f", ".", "read", "(", ")", "setting_dict", "=", "{", "'settings_default'", ":", "os", ".", "path", ".", "join", "(", "EVENNIA_LIB", ",", "'settings_default.py'", ")", ",", "'servername'", ":", "(", "'\"%s\"'", "%", "GAMEDIR", ".", "rsplit", "(", "os", ".", "path", ".", "sep", ",", "1", ")", "[", "1", "]", ".", "capitalize", "(", ")", ")", ",", "'secret_key'", ":", "(", "\"'%s'\"", "%", "create_secret_key", "(", ")", ")", "}", "settings_string", "=", "settings_string", ".", "format", "(", "**", "setting_dict", ")", "with", "open", "(", "settings_path", ",", "'w'", ")", "as", "f", ":", "f", ".", "write", "(", "settings_string", ")" ]
uses the template settings file to build a working settings file .
train
false
23,680
@utils.arg('--tenant', metavar='<tenant-id>', default=None, help=_('ID of tenant to list the quotas for.')) @utils.arg('--user', metavar='<user-id>', default=None, help=_('ID of user to list the quotas for.')) @utils.arg('--detail', action='store_true', default=False, help=_('Show detailed info (limit, reserved, in-use).')) def do_quota_show(cs, args): if args.tenant: project_id = args.tenant elif isinstance(cs.client, client.SessionClient): auth = cs.client.auth project_id = auth.get_auth_ref(cs.client.session).project_id else: project_id = cs.client.tenant_id _quota_show(cs.quotas.get(project_id, user_id=args.user, detail=args.detail))
[ "@", "utils", ".", "arg", "(", "'--tenant'", ",", "metavar", "=", "'<tenant-id>'", ",", "default", "=", "None", ",", "help", "=", "_", "(", "'ID of tenant to list the quotas for.'", ")", ")", "@", "utils", ".", "arg", "(", "'--user'", ",", "metavar", "=", "'<user-id>'", ",", "default", "=", "None", ",", "help", "=", "_", "(", "'ID of user to list the quotas for.'", ")", ")", "@", "utils", ".", "arg", "(", "'--detail'", ",", "action", "=", "'store_true'", ",", "default", "=", "False", ",", "help", "=", "_", "(", "'Show detailed info (limit, reserved, in-use).'", ")", ")", "def", "do_quota_show", "(", "cs", ",", "args", ")", ":", "if", "args", ".", "tenant", ":", "project_id", "=", "args", ".", "tenant", "elif", "isinstance", "(", "cs", ".", "client", ",", "client", ".", "SessionClient", ")", ":", "auth", "=", "cs", ".", "client", ".", "auth", "project_id", "=", "auth", ".", "get_auth_ref", "(", "cs", ".", "client", ".", "session", ")", ".", "project_id", "else", ":", "project_id", "=", "cs", ".", "client", ".", "tenant_id", "_quota_show", "(", "cs", ".", "quotas", ".", "get", "(", "project_id", ",", "user_id", "=", "args", ".", "user", ",", "detail", "=", "args", ".", "detail", ")", ")" ]
list the quotas for a tenant/user .
train
false
23,681
@pytest.fixture() def pg_xlog(tmpdir, monkeypatch): monkeypatch.chdir(tmpdir) return PgXlog(tmpdir)
[ "@", "pytest", ".", "fixture", "(", ")", "def", "pg_xlog", "(", "tmpdir", ",", "monkeypatch", ")", ":", "monkeypatch", ".", "chdir", "(", "tmpdir", ")", "return", "PgXlog", "(", "tmpdir", ")" ]
set up xlog utility functions and change directories .
train
false
23,682
def _datetime_to_rfc3339(value, ignore_zone=True): if ((not ignore_zone) and (value.tzinfo is not None)): value = (value.replace(tzinfo=None) - value.utcoffset()) return value.strftime(_RFC3339_MICROS)
[ "def", "_datetime_to_rfc3339", "(", "value", ",", "ignore_zone", "=", "True", ")", ":", "if", "(", "(", "not", "ignore_zone", ")", "and", "(", "value", ".", "tzinfo", "is", "not", "None", ")", ")", ":", "value", "=", "(", "value", ".", "replace", "(", "tzinfo", "=", "None", ")", "-", "value", ".", "utcoffset", "(", ")", ")", "return", "value", ".", "strftime", "(", "_RFC3339_MICROS", ")" ]
convert a timestamp to a string .
train
true
23,683
def _unconstrain_sv_less_than_one(constrained, order=None, k_endog=None): from scipy import linalg unconstrained = [] if (order is None): order = len(constrained) if (k_endog is None): k_endog = constrained[0].shape[0] eye = np.eye(k_endog) for i in range(order): P = constrained[i] (B_inv, lower) = linalg.cho_factor((eye - np.dot(P, P.T)), lower=True) unconstrained.append(linalg.solve_triangular(B_inv, P, lower=lower)) return unconstrained
[ "def", "_unconstrain_sv_less_than_one", "(", "constrained", ",", "order", "=", "None", ",", "k_endog", "=", "None", ")", ":", "from", "scipy", "import", "linalg", "unconstrained", "=", "[", "]", "if", "(", "order", "is", "None", ")", ":", "order", "=", "len", "(", "constrained", ")", "if", "(", "k_endog", "is", "None", ")", ":", "k_endog", "=", "constrained", "[", "0", "]", ".", "shape", "[", "0", "]", "eye", "=", "np", ".", "eye", "(", "k_endog", ")", "for", "i", "in", "range", "(", "order", ")", ":", "P", "=", "constrained", "[", "i", "]", "(", "B_inv", ",", "lower", ")", "=", "linalg", ".", "cho_factor", "(", "(", "eye", "-", "np", ".", "dot", "(", "P", ",", "P", ".", "T", ")", ")", ",", "lower", "=", "True", ")", "unconstrained", ".", "append", "(", "linalg", ".", "solve_triangular", "(", "B_inv", ",", "P", ",", "lower", "=", "lower", ")", ")", "return", "unconstrained" ]
transform matrices with singular values less than one to arbitrary matrices .
train
false
23,684
@app.route('/calendars/<public_id>', methods=['GET']) def calendar_read_api(public_id): valid_public_id(public_id) try: calendar = g.db_session.query(Calendar).filter((Calendar.public_id == public_id), (Calendar.namespace_id == g.namespace.id)).one() except NoResultFound: raise NotFoundError("Couldn't find calendar {0}".format(public_id)) return g.encoder.jsonify(calendar)
[ "@", "app", ".", "route", "(", "'/calendars/<public_id>'", ",", "methods", "=", "[", "'GET'", "]", ")", "def", "calendar_read_api", "(", "public_id", ")", ":", "valid_public_id", "(", "public_id", ")", "try", ":", "calendar", "=", "g", ".", "db_session", ".", "query", "(", "Calendar", ")", ".", "filter", "(", "(", "Calendar", ".", "public_id", "==", "public_id", ")", ",", "(", "Calendar", ".", "namespace_id", "==", "g", ".", "namespace", ".", "id", ")", ")", ".", "one", "(", ")", "except", "NoResultFound", ":", "raise", "NotFoundError", "(", "\"Couldn't find calendar {0}\"", ".", "format", "(", "public_id", ")", ")", "return", "g", ".", "encoder", ".", "jsonify", "(", "calendar", ")" ]
get all data for an existing calendar .
train
false
23,685
def labels_from_header(header_name, header_value): if header_value: labels = parse_labels_string(header_name, header_value) else: labels = set() return labels
[ "def", "labels_from_header", "(", "header_name", ",", "header_value", ")", ":", "if", "header_value", ":", "labels", "=", "parse_labels_string", "(", "header_name", ",", "header_value", ")", "else", ":", "labels", "=", "set", "(", ")", "return", "labels" ]
helper that builds label set from the corresponding header value .
train
false
23,686
def load_rt(loc='./data/'): (pos, neg) = ([], []) with open((loc + 'rt-polarity.pos'), 'rb') as f: for line in f: pos.append(line.decode('latin-1').strip()) with open((loc + 'rt-polarity.neg'), 'rb') as f: for line in f: neg.append(line.decode('latin-1').strip()) return (pos, neg)
[ "def", "load_rt", "(", "loc", "=", "'./data/'", ")", ":", "(", "pos", ",", "neg", ")", "=", "(", "[", "]", ",", "[", "]", ")", "with", "open", "(", "(", "loc", "+", "'rt-polarity.pos'", ")", ",", "'rb'", ")", "as", "f", ":", "for", "line", "in", "f", ":", "pos", ".", "append", "(", "line", ".", "decode", "(", "'latin-1'", ")", ".", "strip", "(", ")", ")", "with", "open", "(", "(", "loc", "+", "'rt-polarity.neg'", ")", ",", "'rb'", ")", "as", "f", ":", "for", "line", "in", "f", ":", "neg", ".", "append", "(", "line", ".", "decode", "(", "'latin-1'", ")", ".", "strip", "(", ")", ")", "return", "(", "pos", ",", "neg", ")" ]
load the mr dataset .
train
false
23,688
def semilinearPrime(x): try: shape = x.shape x.flatten() x = x.tolist() except AttributeError: shape = (1, len(x)) def f(val): if (val < 0): return safeExp(val) else: return 1.0 return array(list(map(f, x))).reshape(shape)
[ "def", "semilinearPrime", "(", "x", ")", ":", "try", ":", "shape", "=", "x", ".", "shape", "x", ".", "flatten", "(", ")", "x", "=", "x", ".", "tolist", "(", ")", "except", "AttributeError", ":", "shape", "=", "(", "1", ",", "len", "(", "x", ")", ")", "def", "f", "(", "val", ")", ":", "if", "(", "val", "<", "0", ")", ":", "return", "safeExp", "(", "val", ")", "else", ":", "return", "1.0", "return", "array", "(", "list", "(", "map", "(", "f", ",", "x", ")", ")", ")", ".", "reshape", "(", "shape", ")" ]
this function is the first derivative of the semilinear function .
train
false
23,689
def mergeOrderings(orderings, seen=None): if (seen is None): seen = {} result = [] orderings.reverse() for ordering in orderings: ordering = list(ordering) ordering.reverse() for o in ordering: if (o not in seen): seen[o] = 1 result.append(o) result.reverse() return result
[ "def", "mergeOrderings", "(", "orderings", ",", "seen", "=", "None", ")", ":", "if", "(", "seen", "is", "None", ")", ":", "seen", "=", "{", "}", "result", "=", "[", "]", "orderings", ".", "reverse", "(", ")", "for", "ordering", "in", "orderings", ":", "ordering", "=", "list", "(", "ordering", ")", "ordering", ".", "reverse", "(", ")", "for", "o", "in", "ordering", ":", "if", "(", "o", "not", "in", "seen", ")", ":", "seen", "[", "o", "]", "=", "1", "result", ".", "append", "(", "o", ")", "result", ".", "reverse", "(", ")", "return", "result" ]
merge multiple orderings so that within-ordering order is preserved orderings are constrained in such a way that if an object appears in two or more orderings .
train
false
23,690
def wrap_output(output, encoding): return codecs.getwriter(encoding)((output.buffer if hasattr(output, u'buffer') else output))
[ "def", "wrap_output", "(", "output", ",", "encoding", ")", ":", "return", "codecs", ".", "getwriter", "(", "encoding", ")", "(", "(", "output", ".", "buffer", "if", "hasattr", "(", "output", ",", "u'buffer'", ")", "else", "output", ")", ")" ]
return output with specified encoding .
train
true
23,691
def ordered_intersect(*sets): common = frozenset.intersection(*map(frozenset, sets)) return (x for x in unique(concat(sets)) if (x in common))
[ "def", "ordered_intersect", "(", "*", "sets", ")", ":", "common", "=", "frozenset", ".", "intersection", "(", "*", "map", "(", "frozenset", ",", "sets", ")", ")", "return", "(", "x", "for", "x", "in", "unique", "(", "concat", "(", "sets", ")", ")", "if", "(", "x", "in", "common", ")", ")" ]
set intersection of two sequences that preserves order .
train
false
23,693
def determine_64_bit_int(): try: try: import ctypes except ImportError: raise ValueError() if (ctypes.sizeof(ctypes.c_longlong) == 8): return u'long long int' elif (ctypes.sizeof(ctypes.c_long) == 8): return u'long int' elif (ctypes.sizeof(ctypes.c_int) == 8): return u'int' else: raise ValueError() except ValueError: return u'long long int'
[ "def", "determine_64_bit_int", "(", ")", ":", "try", ":", "try", ":", "import", "ctypes", "except", "ImportError", ":", "raise", "ValueError", "(", ")", "if", "(", "ctypes", ".", "sizeof", "(", "ctypes", ".", "c_longlong", ")", "==", "8", ")", ":", "return", "u'long long int'", "elif", "(", "ctypes", ".", "sizeof", "(", "ctypes", ".", "c_long", ")", "==", "8", ")", ":", "return", "u'long int'", "elif", "(", "ctypes", ".", "sizeof", "(", "ctypes", ".", "c_int", ")", "==", "8", ")", ":", "return", "u'int'", "else", ":", "raise", "ValueError", "(", ")", "except", "ValueError", ":", "return", "u'long long int'" ]
the only configuration parameter needed at compile-time is how to specify a 64-bit signed integer .
train
false
23,694
def get_predicted_pageviews(srs, location=None): (srs, is_single) = tup(srs, ret_is_single=True) sr_names = [sr.name for sr in srs] default_srids = LocalizedDefaultSubreddits.get_global_defaults() if location: no_location = Location(None) r = LocationPromoMetrics.get(DefaultSR, [no_location, location]) location_pageviews = r[(DefaultSR, location)] all_pageviews = r[(DefaultSR, no_location)] if all_pageviews: location_factor = (float(location_pageviews) / float(all_pageviews)) else: location_factor = 0.0 else: location_factor = 1.0 daily_inventory = PromoMetrics.get(MIN_DAILY_CASS_KEY, sr_names=sr_names) ret = {} for sr in srs: if ((not isinstance(sr, FakeSubreddit)) and (sr._id in default_srids)): default_factor = DEFAULT_INVENTORY_FACTOR else: default_factor = INVENTORY_FACTOR base_pageviews = daily_inventory.get(sr.name, 0) ret[sr.name] = int(((base_pageviews * default_factor) * location_factor)) if is_single: return ret[srs[0].name] else: return ret
[ "def", "get_predicted_pageviews", "(", "srs", ",", "location", "=", "None", ")", ":", "(", "srs", ",", "is_single", ")", "=", "tup", "(", "srs", ",", "ret_is_single", "=", "True", ")", "sr_names", "=", "[", "sr", ".", "name", "for", "sr", "in", "srs", "]", "default_srids", "=", "LocalizedDefaultSubreddits", ".", "get_global_defaults", "(", ")", "if", "location", ":", "no_location", "=", "Location", "(", "None", ")", "r", "=", "LocationPromoMetrics", ".", "get", "(", "DefaultSR", ",", "[", "no_location", ",", "location", "]", ")", "location_pageviews", "=", "r", "[", "(", "DefaultSR", ",", "location", ")", "]", "all_pageviews", "=", "r", "[", "(", "DefaultSR", ",", "no_location", ")", "]", "if", "all_pageviews", ":", "location_factor", "=", "(", "float", "(", "location_pageviews", ")", "/", "float", "(", "all_pageviews", ")", ")", "else", ":", "location_factor", "=", "0.0", "else", ":", "location_factor", "=", "1.0", "daily_inventory", "=", "PromoMetrics", ".", "get", "(", "MIN_DAILY_CASS_KEY", ",", "sr_names", "=", "sr_names", ")", "ret", "=", "{", "}", "for", "sr", "in", "srs", ":", "if", "(", "(", "not", "isinstance", "(", "sr", ",", "FakeSubreddit", ")", ")", "and", "(", "sr", ".", "_id", "in", "default_srids", ")", ")", ":", "default_factor", "=", "DEFAULT_INVENTORY_FACTOR", "else", ":", "default_factor", "=", "INVENTORY_FACTOR", "base_pageviews", "=", "daily_inventory", ".", "get", "(", "sr", ".", "name", ",", "0", ")", "ret", "[", "sr", ".", "name", "]", "=", "int", "(", "(", "(", "base_pageviews", "*", "default_factor", ")", "*", "location_factor", ")", ")", "if", "is_single", ":", "return", "ret", "[", "srs", "[", "0", "]", ".", "name", "]", "else", ":", "return", "ret" ]
return predicted number of pageviews for sponsored headlines .
train
false
23,696
def update_node_links(designated_node, target_nodes, description): if (len(target_nodes) == 0): logger.info('No target nodes specified - no node links will be added!') else: logger.info('Repopulating {} with latest {} nodes.'.format(designated_node._id, description)) user = designated_node.creator auth = Auth(user) for pointer in designated_node.nodes_pointer: designated_node.rm_pointer(pointer, auth) for node in target_nodes: designated_node.add_pointer(node, auth, save=True) logger.info('Added node link {} to {}'.format(node, designated_node))
[ "def", "update_node_links", "(", "designated_node", ",", "target_nodes", ",", "description", ")", ":", "if", "(", "len", "(", "target_nodes", ")", "==", "0", ")", ":", "logger", ".", "info", "(", "'No target nodes specified - no node links will be added!'", ")", "else", ":", "logger", ".", "info", "(", "'Repopulating {} with latest {} nodes.'", ".", "format", "(", "designated_node", ".", "_id", ",", "description", ")", ")", "user", "=", "designated_node", ".", "creator", "auth", "=", "Auth", "(", "user", ")", "for", "pointer", "in", "designated_node", ".", "nodes_pointer", ":", "designated_node", ".", "rm_pointer", "(", "pointer", ",", "auth", ")", "for", "node", "in", "target_nodes", ":", "designated_node", ".", "add_pointer", "(", "node", ",", "auth", ",", "save", "=", "True", ")", "logger", ".", "info", "(", "'Added node link {} to {}'", ".", "format", "(", "node", ",", "designated_node", ")", ")" ]
takes designated node .
train
false
23,698
def compressString(s): gzip.time = FakeTime() zbuf = BytesIO() zfile = gzip.GzipFile(mode='wb', compresslevel=9, fileobj=zbuf) zfile.write(s) zfile.close() return zbuf.getvalue()
[ "def", "compressString", "(", "s", ")", ":", "gzip", ".", "time", "=", "FakeTime", "(", ")", "zbuf", "=", "BytesIO", "(", ")", "zfile", "=", "gzip", ".", "GzipFile", "(", "mode", "=", "'wb'", ",", "compresslevel", "=", "9", ",", "fileobj", "=", "zbuf", ")", "zfile", ".", "write", "(", "s", ")", "zfile", ".", "close", "(", ")", "return", "zbuf", ".", "getvalue", "(", ")" ]
gzip a given string .
train
false
23,699
def rpn_generate(queue=None, imdb_name=None, rpn_model_path=None, cfg=None, rpn_test_prototxt=None): cfg.TEST.RPN_PRE_NMS_TOP_N = (-1) cfg.TEST.RPN_POST_NMS_TOP_N = 2000 print 'RPN model: {}'.format(rpn_model_path) print 'Using config:' pprint.pprint(cfg) import caffe _init_caffe(cfg) imdb = get_imdb(imdb_name) print 'Loaded dataset `{:s}` for proposal generation'.format(imdb.name) rpn_net = caffe.Net(rpn_test_prototxt, rpn_model_path, caffe.TEST) output_dir = get_output_dir(imdb) print 'Output will be saved to `{:s}`'.format(output_dir) rpn_proposals = imdb_proposals(rpn_net, imdb) rpn_net_name = os.path.splitext(os.path.basename(rpn_model_path))[0] rpn_proposals_path = os.path.join(output_dir, (rpn_net_name + '_proposals.pkl')) with open(rpn_proposals_path, 'wb') as f: cPickle.dump(rpn_proposals, f, cPickle.HIGHEST_PROTOCOL) print 'Wrote RPN proposals to {}'.format(rpn_proposals_path) queue.put({'proposal_path': rpn_proposals_path})
[ "def", "rpn_generate", "(", "queue", "=", "None", ",", "imdb_name", "=", "None", ",", "rpn_model_path", "=", "None", ",", "cfg", "=", "None", ",", "rpn_test_prototxt", "=", "None", ")", ":", "cfg", ".", "TEST", ".", "RPN_PRE_NMS_TOP_N", "=", "(", "-", "1", ")", "cfg", ".", "TEST", ".", "RPN_POST_NMS_TOP_N", "=", "2000", "print", "'RPN model: {}'", ".", "format", "(", "rpn_model_path", ")", "print", "'Using config:'", "pprint", ".", "pprint", "(", "cfg", ")", "import", "caffe", "_init_caffe", "(", "cfg", ")", "imdb", "=", "get_imdb", "(", "imdb_name", ")", "print", "'Loaded dataset `{:s}` for proposal generation'", ".", "format", "(", "imdb", ".", "name", ")", "rpn_net", "=", "caffe", ".", "Net", "(", "rpn_test_prototxt", ",", "rpn_model_path", ",", "caffe", ".", "TEST", ")", "output_dir", "=", "get_output_dir", "(", "imdb", ")", "print", "'Output will be saved to `{:s}`'", ".", "format", "(", "output_dir", ")", "rpn_proposals", "=", "imdb_proposals", "(", "rpn_net", ",", "imdb", ")", "rpn_net_name", "=", "os", ".", "path", ".", "splitext", "(", "os", ".", "path", ".", "basename", "(", "rpn_model_path", ")", ")", "[", "0", "]", "rpn_proposals_path", "=", "os", ".", "path", ".", "join", "(", "output_dir", ",", "(", "rpn_net_name", "+", "'_proposals.pkl'", ")", ")", "with", "open", "(", "rpn_proposals_path", ",", "'wb'", ")", "as", "f", ":", "cPickle", ".", "dump", "(", "rpn_proposals", ",", "f", ",", "cPickle", ".", "HIGHEST_PROTOCOL", ")", "print", "'Wrote RPN proposals to {}'", ".", "format", "(", "rpn_proposals_path", ")", "queue", ".", "put", "(", "{", "'proposal_path'", ":", "rpn_proposals_path", "}", ")" ]
use a trained rpn to generate proposals .
train
false
23,700
def clean_search_unit(pk, lang): if appsettings.OFFLOAD_INDEXING: add_index_update(pk, False, True, lang) else: delete_search_unit(pk, lang)
[ "def", "clean_search_unit", "(", "pk", ",", "lang", ")", ":", "if", "appsettings", ".", "OFFLOAD_INDEXING", ":", "add_index_update", "(", "pk", ",", "False", ",", "True", ",", "lang", ")", "else", ":", "delete_search_unit", "(", "pk", ",", "lang", ")" ]
cleanups search index on unit deletion .
train
false
23,701
def topic_sibling_documents_link(obj): if (not obj.parent_topic): return '' count = obj.parent_topic.children.count() if (not count): return '' link = ('%s?%s' % (reverse('admin:wiki_document_changelist', args=[]), ('parent_topic__exact=%s' % obj.parent_topic.id))) what = (((count == 1) and 'sibling') or 'siblings') return ('<a href="%s">%s&nbsp;%s</a>' % (link, count, what))
[ "def", "topic_sibling_documents_link", "(", "obj", ")", ":", "if", "(", "not", "obj", ".", "parent_topic", ")", ":", "return", "''", "count", "=", "obj", ".", "parent_topic", ".", "children", ".", "count", "(", ")", "if", "(", "not", "count", ")", ":", "return", "''", "link", "=", "(", "'%s?%s'", "%", "(", "reverse", "(", "'admin:wiki_document_changelist'", ",", "args", "=", "[", "]", ")", ",", "(", "'parent_topic__exact=%s'", "%", "obj", ".", "parent_topic", ".", "id", ")", ")", ")", "what", "=", "(", "(", "(", "count", "==", "1", ")", "and", "'sibling'", ")", "or", "'siblings'", ")", "return", "(", "'<a href=\"%s\">%s&nbsp;%s</a>'", "%", "(", "link", ",", "count", ",", "what", ")", ")" ]
html link to a list of sibling documents .
train
false
23,702
def build_block_parser(md_instance, **kwargs): parser = BlockParser(md_instance) parser.blockprocessors[u'empty'] = EmptyBlockProcessor(parser) parser.blockprocessors[u'indent'] = ListIndentProcessor(parser) parser.blockprocessors[u'code'] = CodeBlockProcessor(parser) parser.blockprocessors[u'hashheader'] = HashHeaderProcessor(parser) parser.blockprocessors[u'setextheader'] = SetextHeaderProcessor(parser) parser.blockprocessors[u'hr'] = HRProcessor(parser) parser.blockprocessors[u'olist'] = OListProcessor(parser) parser.blockprocessors[u'ulist'] = UListProcessor(parser) parser.blockprocessors[u'quote'] = BlockQuoteProcessor(parser) parser.blockprocessors[u'paragraph'] = ParagraphProcessor(parser) return parser
[ "def", "build_block_parser", "(", "md_instance", ",", "**", "kwargs", ")", ":", "parser", "=", "BlockParser", "(", "md_instance", ")", "parser", ".", "blockprocessors", "[", "u'empty'", "]", "=", "EmptyBlockProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'indent'", "]", "=", "ListIndentProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'code'", "]", "=", "CodeBlockProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'hashheader'", "]", "=", "HashHeaderProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'setextheader'", "]", "=", "SetextHeaderProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'hr'", "]", "=", "HRProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'olist'", "]", "=", "OListProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'ulist'", "]", "=", "UListProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'quote'", "]", "=", "BlockQuoteProcessor", "(", "parser", ")", "parser", ".", "blockprocessors", "[", "u'paragraph'", "]", "=", "ParagraphProcessor", "(", "parser", ")", "return", "parser" ]
build the default block parser used by markdown .
train
false
23,705
def _wrap_lines(msg): lines = msg.splitlines() fixed_l = [] for line in lines: fixed_l.append(textwrap.fill(line, 80, break_long_words=False, break_on_hyphens=False)) return os.linesep.join(fixed_l)
[ "def", "_wrap_lines", "(", "msg", ")", ":", "lines", "=", "msg", ".", "splitlines", "(", ")", "fixed_l", "=", "[", "]", "for", "line", "in", "lines", ":", "fixed_l", ".", "append", "(", "textwrap", ".", "fill", "(", "line", ",", "80", ",", "break_long_words", "=", "False", ",", "break_on_hyphens", "=", "False", ")", ")", "return", "os", ".", "linesep", ".", "join", "(", "fixed_l", ")" ]
format lines nicely to 80 chars .
train
false
23,706
def add_contributor_json(user, current_user=None): if current_user: n_projects_in_common = current_user.n_projects_in_common(user) else: n_projects_in_common = 0 current_employment = None education = None if user.jobs: current_employment = user.jobs[0]['institution'] if user.schools: education = user.schools[0]['institution'] return {'fullname': user.fullname, 'email': user.username, 'id': user._primary_key, 'employment': current_employment, 'education': education, 'n_projects_in_common': n_projects_in_common, 'registered': user.is_registered, 'active': user.is_active, 'gravatar_url': gravatar(user, use_ssl=True, size=settings.PROFILE_IMAGE_MEDIUM), 'profile_url': user.profile_url}
[ "def", "add_contributor_json", "(", "user", ",", "current_user", "=", "None", ")", ":", "if", "current_user", ":", "n_projects_in_common", "=", "current_user", ".", "n_projects_in_common", "(", "user", ")", "else", ":", "n_projects_in_common", "=", "0", "current_employment", "=", "None", "education", "=", "None", "if", "user", ".", "jobs", ":", "current_employment", "=", "user", ".", "jobs", "[", "0", "]", "[", "'institution'", "]", "if", "user", ".", "schools", ":", "education", "=", "user", ".", "schools", "[", "0", "]", "[", "'institution'", "]", "return", "{", "'fullname'", ":", "user", ".", "fullname", ",", "'email'", ":", "user", ".", "username", ",", "'id'", ":", "user", ".", "_primary_key", ",", "'employment'", ":", "current_employment", ",", "'education'", ":", "education", ",", "'n_projects_in_common'", ":", "n_projects_in_common", ",", "'registered'", ":", "user", ".", "is_registered", ",", "'active'", ":", "user", ".", "is_active", ",", "'gravatar_url'", ":", "gravatar", "(", "user", ",", "use_ssl", "=", "True", ",", "size", "=", "settings", ".", "PROFILE_IMAGE_MEDIUM", ")", ",", "'profile_url'", ":", "user", ".", "profile_url", "}" ]
generate a dictionary representation of a user .
train
false
23,707
def get_messages(request): return getattr(request, '_messages', [])
[ "def", "get_messages", "(", "request", ")", ":", "return", "getattr", "(", "request", ",", "'_messages'", ",", "[", "]", ")" ]
returns the message storage on the request if it exists .
train
false
23,708
def hot(): rc(u'image', cmap=u'hot') im = gci() if (im is not None): im.set_cmap(cm.hot)
[ "def", "hot", "(", ")", ":", "rc", "(", "u'image'", ",", "cmap", "=", "u'hot'", ")", "im", "=", "gci", "(", ")", "if", "(", "im", "is", "not", "None", ")", ":", "im", ".", "set_cmap", "(", "cm", ".", "hot", ")" ]
returns a list of hit terms via google trends .
train
false
23,709
def regressionWrapper(model, modelType, testSample): if ((modelType == 'svm') or (modelType == 'randomforest')): return model.predict(testSample.reshape(1, (-1)))[0] return None
[ "def", "regressionWrapper", "(", "model", ",", "modelType", ",", "testSample", ")", ":", "if", "(", "(", "modelType", "==", "'svm'", ")", "or", "(", "modelType", "==", "'randomforest'", ")", ")", ":", "return", "model", ".", "predict", "(", "testSample", ".", "reshape", "(", "1", ",", "(", "-", "1", ")", ")", ")", "[", "0", "]", "return", "None" ]
this function is used as a wrapper to pattern classification .
train
false
23,711
def _get_tag_int(fid, node, name, id_): tag = find_tag(fid, node, id_) if (tag is None): fid.close() raise ValueError((name + ' tag not found')) return int(tag.data)
[ "def", "_get_tag_int", "(", "fid", ",", "node", ",", "name", ",", "id_", ")", ":", "tag", "=", "find_tag", "(", "fid", ",", "node", ",", "id_", ")", "if", "(", "tag", "is", "None", ")", ":", "fid", ".", "close", "(", ")", "raise", "ValueError", "(", "(", "name", "+", "' tag not found'", ")", ")", "return", "int", "(", "tag", ".", "data", ")" ]
check we have an appropriate tag .
train
false
23,713
def on_off(tag): return ['OFF', 'ON'][tag]
[ "def", "on_off", "(", "tag", ")", ":", "return", "[", "'OFF'", ",", "'ON'", "]", "[", "tag", "]" ]
return an on/off string for a 1/0 input .
train
false
23,714
def typestats(objects=None, shortnames=True): if (objects is None): objects = gc.get_objects() try: if shortnames: typename = _short_typename else: typename = _long_typename stats = {} for o in objects: n = typename(o) stats[n] = (stats.get(n, 0) + 1) return stats finally: del objects
[ "def", "typestats", "(", "objects", "=", "None", ",", "shortnames", "=", "True", ")", ":", "if", "(", "objects", "is", "None", ")", ":", "objects", "=", "gc", ".", "get_objects", "(", ")", "try", ":", "if", "shortnames", ":", "typename", "=", "_short_typename", "else", ":", "typename", "=", "_long_typename", "stats", "=", "{", "}", "for", "o", "in", "objects", ":", "n", "=", "typename", "(", "o", ")", "stats", "[", "n", "]", "=", "(", "stats", ".", "get", "(", "n", ",", "0", ")", "+", "1", ")", "return", "stats", "finally", ":", "del", "objects" ]
count the number of instances for each type tracked by the gc .
train
false
23,715
def remove_session(session): Session.remove_one(session)
[ "def", "remove_session", "(", "session", ")", ":", "Session", ".", "remove_one", "(", "session", ")" ]
remove a session from database .
train
false
23,716
def test_success_junit_xml(test_name, class_name='Results', testcase_name='test_ran'): testsuite = ET.Element('testsuite') testsuite.set('tests', '1') testsuite.set('failures', '0') testsuite.set('time', '1') testsuite.set('errors', '0') testsuite.set('name', test_name) testcase = ET.SubElement(testsuite, 'testcase') testcase.set('name', testcase_name) testcase.set('status', 'run') testcase.set('time', '1') testcase.set('classname', class_name) return ET.tostring(testsuite, encoding='utf-8', method='xml')
[ "def", "test_success_junit_xml", "(", "test_name", ",", "class_name", "=", "'Results'", ",", "testcase_name", "=", "'test_ran'", ")", ":", "testsuite", "=", "ET", ".", "Element", "(", "'testsuite'", ")", "testsuite", ".", "set", "(", "'tests'", ",", "'1'", ")", "testsuite", ".", "set", "(", "'failures'", ",", "'0'", ")", "testsuite", ".", "set", "(", "'time'", ",", "'1'", ")", "testsuite", ".", "set", "(", "'errors'", ",", "'0'", ")", "testsuite", ".", "set", "(", "'name'", ",", "test_name", ")", "testcase", "=", "ET", ".", "SubElement", "(", "testsuite", ",", "'testcase'", ")", "testcase", ".", "set", "(", "'name'", ",", "testcase_name", ")", "testcase", ".", "set", "(", "'status'", ",", "'run'", ")", "testcase", ".", "set", "(", "'time'", ",", "'1'", ")", "testcase", ".", "set", "(", "'classname'", ",", "class_name", ")", "return", "ET", ".", "tostring", "(", "testsuite", ",", "encoding", "=", "'utf-8'", ",", "method", "=", "'xml'", ")" ]
generate junit xml file for a unary test suite where the test succeeded .
train
false
23,717
def varexp(line): ip = get_ipython() (funcname, name) = line.split() import spyder.pyplot __fig__ = spyder.pyplot.figure() __items__ = getattr(spyder.pyplot, funcname[2:])(ip.user_ns[name]) spyder.pyplot.show() del __fig__, __items__
[ "def", "varexp", "(", "line", ")", ":", "ip", "=", "get_ipython", "(", ")", "(", "funcname", ",", "name", ")", "=", "line", ".", "split", "(", ")", "import", "spyder", ".", "pyplot", "__fig__", "=", "spyder", ".", "pyplot", ".", "figure", "(", ")", "__items__", "=", "getattr", "(", "spyder", ".", "pyplot", ",", "funcname", "[", "2", ":", "]", ")", "(", "ip", ".", "user_ns", "[", "name", "]", ")", "spyder", ".", "pyplot", ".", "show", "(", ")", "del", "__fig__", ",", "__items__" ]
spyders variable explorer magic used to generate plots .
train
true
23,718
def test_initial_column_fill_values(): class TestHeader(ascii.BasicHeader, ): def _set_cols_from_names(self): self.cols = [ascii.Column(name=x) for x in self.names] for col in self.cols: col.fill_values = {'--': '0'} class Tester(ascii.Basic, ): header_class = TestHeader reader = ascii.get_reader(Reader=Tester) assert (reader.read('# Column definition is the first uncommented line\n# Default delimiter is the space character.\na b c\n# Data starts after the header column definition, blank lines ignored\n-- 2 3\n4 5 6 ')['a'][0] is np.ma.masked)
[ "def", "test_initial_column_fill_values", "(", ")", ":", "class", "TestHeader", "(", "ascii", ".", "BasicHeader", ",", ")", ":", "def", "_set_cols_from_names", "(", "self", ")", ":", "self", ".", "cols", "=", "[", "ascii", ".", "Column", "(", "name", "=", "x", ")", "for", "x", "in", "self", ".", "names", "]", "for", "col", "in", "self", ".", "cols", ":", "col", ".", "fill_values", "=", "{", "'--'", ":", "'0'", "}", "class", "Tester", "(", "ascii", ".", "Basic", ",", ")", ":", "header_class", "=", "TestHeader", "reader", "=", "ascii", ".", "get_reader", "(", "Reader", "=", "Tester", ")", "assert", "(", "reader", ".", "read", "(", "'# Column definition is the first uncommented line\\n# Default delimiter is the space character.\\na b c\\n# Data starts after the header column definition, blank lines ignored\\n-- 2 3\\n4 5 6 '", ")", "[", "'a'", "]", "[", "0", "]", "is", "np", ".", "ma", ".", "masked", ")" ]
regression test for #5336 .
train
false
23,719
def get_stripped_lines(string, ignore_lines_starting_with=''): string = unicode(string) lines = [unicode(l.strip()) for l in string.splitlines()] if ignore_lines_starting_with: filter_func = (lambda x: (x and (not x.startswith(ignore_lines_starting_with)))) else: filter_func = (lambda x: x) lines = filter(filter_func, lines) return lines
[ "def", "get_stripped_lines", "(", "string", ",", "ignore_lines_starting_with", "=", "''", ")", ":", "string", "=", "unicode", "(", "string", ")", "lines", "=", "[", "unicode", "(", "l", ".", "strip", "(", ")", ")", "for", "l", "in", "string", ".", "splitlines", "(", ")", "]", "if", "ignore_lines_starting_with", ":", "filter_func", "=", "(", "lambda", "x", ":", "(", "x", "and", "(", "not", "x", ".", "startswith", "(", "ignore_lines_starting_with", ")", ")", ")", ")", "else", ":", "filter_func", "=", "(", "lambda", "x", ":", "x", ")", "lines", "=", "filter", "(", "filter_func", ",", "lines", ")", "return", "lines" ]
split lines at newline char .
train
false
23,720
def set_selectors(facts): deployment_type = facts['common']['deployment_type'] if (deployment_type == 'online'): selector = 'type=infra' else: selector = 'region=infra' if ('hosted' not in facts): facts['hosted'] = {} if ('router' not in facts['hosted']): facts['hosted']['router'] = {} if (('selector' not in facts['hosted']['router']) or (facts['hosted']['router']['selector'] in [None, 'None'])): facts['hosted']['router']['selector'] = selector if ('registry' not in facts['hosted']): facts['hosted']['registry'] = {} if (('selector' not in facts['hosted']['registry']) or (facts['hosted']['registry']['selector'] in [None, 'None'])): facts['hosted']['registry']['selector'] = selector if ('metrics' not in facts['hosted']): facts['hosted']['metrics'] = {} if (('selector' not in facts['hosted']['metrics']) or (facts['hosted']['metrics']['selector'] in [None, 'None'])): facts['hosted']['metrics']['selector'] = None if ('logging' not in facts['hosted']): facts['hosted']['logging'] = {} if (('selector' not in facts['hosted']['logging']) or (facts['hosted']['logging']['selector'] in [None, 'None'])): facts['hosted']['logging']['selector'] = None return facts
[ "def", "set_selectors", "(", "facts", ")", ":", "deployment_type", "=", "facts", "[", "'common'", "]", "[", "'deployment_type'", "]", "if", "(", "deployment_type", "==", "'online'", ")", ":", "selector", "=", "'type=infra'", "else", ":", "selector", "=", "'region=infra'", "if", "(", "'hosted'", "not", "in", "facts", ")", ":", "facts", "[", "'hosted'", "]", "=", "{", "}", "if", "(", "'router'", "not", "in", "facts", "[", "'hosted'", "]", ")", ":", "facts", "[", "'hosted'", "]", "[", "'router'", "]", "=", "{", "}", "if", "(", "(", "'selector'", "not", "in", "facts", "[", "'hosted'", "]", "[", "'router'", "]", ")", "or", "(", "facts", "[", "'hosted'", "]", "[", "'router'", "]", "[", "'selector'", "]", "in", "[", "None", ",", "'None'", "]", ")", ")", ":", "facts", "[", "'hosted'", "]", "[", "'router'", "]", "[", "'selector'", "]", "=", "selector", "if", "(", "'registry'", "not", "in", "facts", "[", "'hosted'", "]", ")", ":", "facts", "[", "'hosted'", "]", "[", "'registry'", "]", "=", "{", "}", "if", "(", "(", "'selector'", "not", "in", "facts", "[", "'hosted'", "]", "[", "'registry'", "]", ")", "or", "(", "facts", "[", "'hosted'", "]", "[", "'registry'", "]", "[", "'selector'", "]", "in", "[", "None", ",", "'None'", "]", ")", ")", ":", "facts", "[", "'hosted'", "]", "[", "'registry'", "]", "[", "'selector'", "]", "=", "selector", "if", "(", "'metrics'", "not", "in", "facts", "[", "'hosted'", "]", ")", ":", "facts", "[", "'hosted'", "]", "[", "'metrics'", "]", "=", "{", "}", "if", "(", "(", "'selector'", "not", "in", "facts", "[", "'hosted'", "]", "[", "'metrics'", "]", ")", "or", "(", "facts", "[", "'hosted'", "]", "[", "'metrics'", "]", "[", "'selector'", "]", "in", "[", "None", ",", "'None'", "]", ")", ")", ":", "facts", "[", "'hosted'", "]", "[", "'metrics'", "]", "[", "'selector'", "]", "=", "None", "if", "(", "'logging'", "not", "in", "facts", "[", "'hosted'", "]", ")", ":", "facts", "[", "'hosted'", "]", "[", "'logging'", "]", "=", "{", "}", "if", "(", "(", "'selector'", "not", "in", "facts", "[", "'hosted'", "]", "[", "'logging'", "]", ")", "or", "(", "facts", "[", "'hosted'", "]", "[", "'logging'", "]", "[", "'selector'", "]", "in", "[", "None", ",", "'None'", "]", ")", ")", ":", "facts", "[", "'hosted'", "]", "[", "'logging'", "]", "[", "'selector'", "]", "=", "None", "return", "facts" ]
set selectors facts if not already present in facts dict args: facts : existing facts returns: dict: the facts dict updated with the generated selectors facts if they were not already present .
train
false
23,721
def read_in(path): with open(path, 'r') as file_fd: return file_fd.read()
[ "def", "read_in", "(", "path", ")", ":", "with", "open", "(", "path", ",", "'r'", ")", "as", "file_fd", ":", "return", "file_fd", ".", "read", "(", ")" ]
read in a file .
train
false
23,723
def get_parent_url(module, xml=None): if hasattr(module, 'xml_attributes'): return module.xml_attributes.get('parent_url', module.xml_attributes.get('parent_sequential_url')) if (xml is not None): create_xml_attributes(module, xml) return get_parent_url(module) return None
[ "def", "get_parent_url", "(", "module", ",", "xml", "=", "None", ")", ":", "if", "hasattr", "(", "module", ",", "'xml_attributes'", ")", ":", "return", "module", ".", "xml_attributes", ".", "get", "(", "'parent_url'", ",", "module", ".", "xml_attributes", ".", "get", "(", "'parent_sequential_url'", ")", ")", "if", "(", "xml", "is", "not", "None", ")", ":", "create_xml_attributes", "(", "module", ",", "xml", ")", "return", "get_parent_url", "(", "module", ")", "return", "None" ]
get the parent_url .
train
false
23,724
@task @log_call @hosts([(u'root@%s' % host) for host in env.hosts]) def secure(new_user=env.user): run(u'apt-get update -q') run(u'apt-get upgrade -y -q') run((u"adduser --gecos '' %s" % new_user)) run((u'usermod -G sudo %s' % new_user)) run(u"sed -i 's:RootLogin yes:RootLogin no:' /etc/ssh/sshd_config") run(u'service ssh restart') print(green((u"Security steps completed. Log in to the server as '%s' from now on." % new_user), bold=True))
[ "@", "task", "@", "log_call", "@", "hosts", "(", "[", "(", "u'root@%s'", "%", "host", ")", "for", "host", "in", "env", ".", "hosts", "]", ")", "def", "secure", "(", "new_user", "=", "env", ".", "user", ")", ":", "run", "(", "u'apt-get update -q'", ")", "run", "(", "u'apt-get upgrade -y -q'", ")", "run", "(", "(", "u\"adduser --gecos '' %s\"", "%", "new_user", ")", ")", "run", "(", "(", "u'usermod -G sudo %s'", "%", "new_user", ")", ")", "run", "(", "u\"sed -i 's:RootLogin yes:RootLogin no:' /etc/ssh/sshd_config\"", ")", "run", "(", "u'service ssh restart'", ")", "print", "(", "green", "(", "(", "u\"Security steps completed. Log in to the server as '%s' from now on.\"", "%", "new_user", ")", ",", "bold", "=", "True", ")", ")" ]
minimal security steps for brand new servers .
train
false
23,726
@utils.singledispatch def typeof_impl(val, c): tp = _typeof_buffer(val, c) if (tp is not None): return tp from . import cffi_utils if cffi_utils.SUPPORTED: if cffi_utils.is_cffi_func(val): return cffi_utils.make_function_type(val) if cffi_utils.is_ffi_instance(val): return types.ffi return getattr(val, '_numba_type_', None)
[ "@", "utils", ".", "singledispatch", "def", "typeof_impl", "(", "val", ",", "c", ")", ":", "tp", "=", "_typeof_buffer", "(", "val", ",", "c", ")", "if", "(", "tp", "is", "not", "None", ")", ":", "return", "tp", "from", ".", "import", "cffi_utils", "if", "cffi_utils", ".", "SUPPORTED", ":", "if", "cffi_utils", ".", "is_cffi_func", "(", "val", ")", ":", "return", "cffi_utils", ".", "make_function_type", "(", "val", ")", "if", "cffi_utils", ".", "is_ffi_instance", "(", "val", ")", ":", "return", "types", ".", "ffi", "return", "getattr", "(", "val", ",", "'_numba_type_'", ",", "None", ")" ]
generic typeof() implementation .
train
false
23,727
def process_envs(attrs=None, where=None): return _osquery_cmd(table='process_envs', attrs=attrs, where=where)
[ "def", "process_envs", "(", "attrs", "=", "None", ",", "where", "=", "None", ")", ":", "return", "_osquery_cmd", "(", "table", "=", "'process_envs'", ",", "attrs", "=", "attrs", ",", "where", "=", "where", ")" ]
return process_envs information from osquery cli example: .
train
false
23,730
def read_ref(refname, repo_dir=None): refs = list_refs(refnames=[refname], repo_dir=repo_dir, limit_to_heads=True) l = tuple(islice(refs, 2)) if l: assert (len(l) == 1) return l[0][1] else: return None
[ "def", "read_ref", "(", "refname", ",", "repo_dir", "=", "None", ")", ":", "refs", "=", "list_refs", "(", "refnames", "=", "[", "refname", "]", ",", "repo_dir", "=", "repo_dir", ",", "limit_to_heads", "=", "True", ")", "l", "=", "tuple", "(", "islice", "(", "refs", ",", "2", ")", ")", "if", "l", ":", "assert", "(", "len", "(", "l", ")", "==", "1", ")", "return", "l", "[", "0", "]", "[", "1", "]", "else", ":", "return", "None" ]
get the commit id of the most recent commit made on a given ref .
train
false
23,731
def test_20news_length_consistency(): try: data = datasets.fetch_20newsgroups(subset='all', download_if_missing=False, shuffle=False) except IOError: raise SkipTest('Download 20 newsgroups to run this test') data = datasets.fetch_20newsgroups(subset='all') assert_equal(len(data['data']), len(data.data)) assert_equal(len(data['target']), len(data.target)) assert_equal(len(data['filenames']), len(data.filenames))
[ "def", "test_20news_length_consistency", "(", ")", ":", "try", ":", "data", "=", "datasets", ".", "fetch_20newsgroups", "(", "subset", "=", "'all'", ",", "download_if_missing", "=", "False", ",", "shuffle", "=", "False", ")", "except", "IOError", ":", "raise", "SkipTest", "(", "'Download 20 newsgroups to run this test'", ")", "data", "=", "datasets", ".", "fetch_20newsgroups", "(", "subset", "=", "'all'", ")", "assert_equal", "(", "len", "(", "data", "[", "'data'", "]", ")", ",", "len", "(", "data", ".", "data", ")", ")", "assert_equal", "(", "len", "(", "data", "[", "'target'", "]", ")", ",", "len", "(", "data", ".", "target", ")", ")", "assert_equal", "(", "len", "(", "data", "[", "'filenames'", "]", ")", ",", "len", "(", "data", ".", "filenames", ")", ")" ]
checks the length consistencies within the bunch this is a non-regression test for a bug present in 0 .
train
false
23,732
def generate_go_binary(target, source, env): return _generate_go_package(target, source, env)
[ "def", "generate_go_binary", "(", "target", ",", "source", ",", "env", ")", ":", "return", "_generate_go_package", "(", "target", ",", "source", ",", "env", ")" ]
generate go command executable .
train
false
23,733
def endpoint_delete(service, profile=None, **connection_args): kstone = auth(profile, **connection_args) endpoint = endpoint_get(service, profile, **connection_args) if ((not endpoint) or ('Error' in endpoint)): return {'Error': 'Could not find any endpoints for the service'} kstone.endpoints.delete(endpoint['id']) endpoint = endpoint_get(service, profile, **connection_args) if ((not endpoint) or ('Error' in endpoint)): return True
[ "def", "endpoint_delete", "(", "service", ",", "profile", "=", "None", ",", "**", "connection_args", ")", ":", "kstone", "=", "auth", "(", "profile", ",", "**", "connection_args", ")", "endpoint", "=", "endpoint_get", "(", "service", ",", "profile", ",", "**", "connection_args", ")", "if", "(", "(", "not", "endpoint", ")", "or", "(", "'Error'", "in", "endpoint", ")", ")", ":", "return", "{", "'Error'", ":", "'Could not find any endpoints for the service'", "}", "kstone", ".", "endpoints", ".", "delete", "(", "endpoint", "[", "'id'", "]", ")", "endpoint", "=", "endpoint_get", "(", "service", ",", "profile", ",", "**", "connection_args", ")", "if", "(", "(", "not", "endpoint", ")", "or", "(", "'Error'", "in", "endpoint", ")", ")", ":", "return", "True" ]
delete endpoints of an openstack service cli examples: .
train
true
23,734
def pg_varchar(size=0): if size: if (not isinstance(size, int)): raise ValueError(('VARCHAR parameter should be an int, got %s' % type(size))) if (size > 0): return ('VARCHAR(%d)' % size) return 'VARCHAR'
[ "def", "pg_varchar", "(", "size", "=", "0", ")", ":", "if", "size", ":", "if", "(", "not", "isinstance", "(", "size", ",", "int", ")", ")", ":", "raise", "ValueError", "(", "(", "'VARCHAR parameter should be an int, got %s'", "%", "type", "(", "size", ")", ")", ")", "if", "(", "size", ">", "0", ")", ":", "return", "(", "'VARCHAR(%d)'", "%", "size", ")", "return", "'VARCHAR'" ]
returns the varchar declaration for the provided size: * if no size return an infinite varchar * otherwise return a varchar(n) :type int size: varchar size .
train
false
23,736
def rehash(): return (win32gui.SendMessageTimeout(win32con.HWND_BROADCAST, win32con.WM_SETTINGCHANGE, 0, 'Environment', 0, 10000)[0] == 1)
[ "def", "rehash", "(", ")", ":", "return", "(", "win32gui", ".", "SendMessageTimeout", "(", "win32con", ".", "HWND_BROADCAST", ",", "win32con", ".", "WM_SETTINGCHANGE", ",", "0", ",", "'Environment'", ",", "0", ",", "10000", ")", "[", "0", "]", "==", "1", ")" ]
run pyenv rehash to update the installed shims .
train
false
23,737
def tx_from_str(txt): import json txt = txt.strip() try: txt.decode('hex') is_hex = True except: is_hex = False if is_hex: return txt tx_dict = json.loads(str(txt)) assert ('hex' in tx_dict.keys()) return tx_dict['hex']
[ "def", "tx_from_str", "(", "txt", ")", ":", "import", "json", "txt", "=", "txt", ".", "strip", "(", ")", "try", ":", "txt", ".", "decode", "(", "'hex'", ")", "is_hex", "=", "True", "except", ":", "is_hex", "=", "False", "if", "is_hex", ":", "return", "txt", "tx_dict", "=", "json", ".", "loads", "(", "str", "(", "txt", ")", ")", "assert", "(", "'hex'", "in", "tx_dict", ".", "keys", "(", ")", ")", "return", "tx_dict", "[", "'hex'", "]" ]
json or raw hexadecimal .
train
false
23,740
def test_cd_home_dir(): homepath = '~/somepath' with cd(homepath): eq_(env.cwd, homepath)
[ "def", "test_cd_home_dir", "(", ")", ":", "homepath", "=", "'~/somepath'", "with", "cd", "(", "homepath", ")", ":", "eq_", "(", "env", ".", "cwd", ",", "homepath", ")" ]
cd() should work with home directories .
train
false
23,741
def check_experiment_dirs(expt_dir): output_subdir = os.path.join(expt_dir, 'output') check_dir(output_subdir) job_subdir = os.path.join(expt_dir, 'jobs') check_dir(job_subdir)
[ "def", "check_experiment_dirs", "(", "expt_dir", ")", ":", "output_subdir", "=", "os", ".", "path", ".", "join", "(", "expt_dir", ",", "'output'", ")", "check_dir", "(", "output_subdir", ")", "job_subdir", "=", "os", ".", "path", ".", "join", "(", "expt_dir", ",", "'jobs'", ")", "check_dir", "(", "job_subdir", ")" ]
make output and jobs sub directories .
train
false
23,742
def get_valid_backend_qos_spec_from_volume_type(volume, volume_type): spec_key_values = get_backend_qos_spec_from_volume_type(volume_type) if (spec_key_values is None): return None validate_qos_spec(spec_key_values) return map_qos_spec(spec_key_values, volume)
[ "def", "get_valid_backend_qos_spec_from_volume_type", "(", "volume", ",", "volume_type", ")", ":", "spec_key_values", "=", "get_backend_qos_spec_from_volume_type", "(", "volume_type", ")", "if", "(", "spec_key_values", "is", "None", ")", ":", "return", "None", "validate_qos_spec", "(", "spec_key_values", ")", "return", "map_qos_spec", "(", "spec_key_values", ",", "volume", ")" ]
given a volume type .
train
false
23,743
def IncrementId(high_id_key): (unused_start, end) = datastore.AllocateIds(high_id_key, max=high_id_key.id()) assert (end >= high_id_key.id())
[ "def", "IncrementId", "(", "high_id_key", ")", ":", "(", "unused_start", ",", "end", ")", "=", "datastore", ".", "AllocateIds", "(", "high_id_key", ",", "max", "=", "high_id_key", ".", "id", "(", ")", ")", "assert", "(", "end", ">=", "high_id_key", ".", "id", "(", ")", ")" ]
increment unique id counter associated with high_id_key beyond high_id_key .
train
false
23,745
def get_zip_class(): class ContextualZipFile(zipfile.ZipFile, ): def __enter__(self): return self def __exit__(self, type, value, traceback): self.close return (zipfile.ZipFile if hasattr(zipfile.ZipFile, '__exit__') else ContextualZipFile)
[ "def", "get_zip_class", "(", ")", ":", "class", "ContextualZipFile", "(", "zipfile", ".", "ZipFile", ",", ")", ":", "def", "__enter__", "(", "self", ")", ":", "return", "self", "def", "__exit__", "(", "self", ",", "type", ",", "value", ",", "traceback", ")", ":", "self", ".", "close", "return", "(", "zipfile", ".", "ZipFile", "if", "hasattr", "(", "zipfile", ".", "ZipFile", ",", "'__exit__'", ")", "else", "ContextualZipFile", ")" ]
supplement zipfile class to support context manager for python 2 .
train
true
23,746
def all_continuous(vars): vars_ = [var for var in vars if (not isinstance(var, pm.model.ObservedRV))] if any([(var.dtype in pm.discrete_types) for var in vars_]): return False else: return True
[ "def", "all_continuous", "(", "vars", ")", ":", "vars_", "=", "[", "var", "for", "var", "in", "vars", "if", "(", "not", "isinstance", "(", "var", ",", "pm", ".", "model", ".", "ObservedRV", ")", ")", "]", "if", "any", "(", "[", "(", "var", ".", "dtype", "in", "pm", ".", "discrete_types", ")", "for", "var", "in", "vars_", "]", ")", ":", "return", "False", "else", ":", "return", "True" ]
check that vars not include discrete variables .
train
false
23,747
def add_container_page_publishing_info(xblock, xblock_info): def safe_get_username(user_id): '\n Guard against bad user_ids, like the infamous "**replace_user**".\n Note that this will ignore our special known IDs (ModuleStoreEnum.UserID).\n We should consider adding special handling for those values.\n\n :param user_id: the user id to get the username of\n :return: username, or None if the user does not exist or user_id is None\n ' if user_id: try: return User.objects.get(id=user_id).username except: pass return None xblock_info['edited_by'] = safe_get_username(xblock.subtree_edited_by) xblock_info['published_by'] = safe_get_username(xblock.published_by) xblock_info['currently_visible_to_students'] = is_currently_visible_to_students(xblock) xblock_info['has_content_group_components'] = has_children_visible_to_specific_content_groups(xblock) if xblock_info['release_date']: xblock_info['release_date_from'] = _get_release_date_from(xblock) if (xblock_info['visibility_state'] == VisibilityState.staff_only): xblock_info['staff_lock_from'] = _get_staff_lock_from(xblock) else: xblock_info['staff_lock_from'] = None
[ "def", "add_container_page_publishing_info", "(", "xblock", ",", "xblock_info", ")", ":", "def", "safe_get_username", "(", "user_id", ")", ":", "if", "user_id", ":", "try", ":", "return", "User", ".", "objects", ".", "get", "(", "id", "=", "user_id", ")", ".", "username", "except", ":", "pass", "return", "None", "xblock_info", "[", "'edited_by'", "]", "=", "safe_get_username", "(", "xblock", ".", "subtree_edited_by", ")", "xblock_info", "[", "'published_by'", "]", "=", "safe_get_username", "(", "xblock", ".", "published_by", ")", "xblock_info", "[", "'currently_visible_to_students'", "]", "=", "is_currently_visible_to_students", "(", "xblock", ")", "xblock_info", "[", "'has_content_group_components'", "]", "=", "has_children_visible_to_specific_content_groups", "(", "xblock", ")", "if", "xblock_info", "[", "'release_date'", "]", ":", "xblock_info", "[", "'release_date_from'", "]", "=", "_get_release_date_from", "(", "xblock", ")", "if", "(", "xblock_info", "[", "'visibility_state'", "]", "==", "VisibilityState", ".", "staff_only", ")", ":", "xblock_info", "[", "'staff_lock_from'", "]", "=", "_get_staff_lock_from", "(", "xblock", ")", "else", ":", "xblock_info", "[", "'staff_lock_from'", "]", "=", "None" ]
adds information about the xblocks publish state to the supplied xblock_info for the container page .
train
false
23,748
def libvlc_video_take_snapshot(p_mi, num, psz_filepath, i_width, i_height): f = (_Cfunctions.get('libvlc_video_take_snapshot', None) or _Cfunction('libvlc_video_take_snapshot', ((1,), (1,), (1,), (1,), (1,)), None, ctypes.c_int, MediaPlayer, ctypes.c_uint, ctypes.c_char_p, ctypes.c_int, ctypes.c_int)) return f(p_mi, num, psz_filepath, i_width, i_height)
[ "def", "libvlc_video_take_snapshot", "(", "p_mi", ",", "num", ",", "psz_filepath", ",", "i_width", ",", "i_height", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_video_take_snapshot'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_video_take_snapshot'", ",", "(", "(", "1", ",", ")", ",", "(", "1", ",", ")", ",", "(", "1", ",", ")", ",", "(", "1", ",", ")", ",", "(", "1", ",", ")", ")", ",", "None", ",", "ctypes", ".", "c_int", ",", "MediaPlayer", ",", "ctypes", ".", "c_uint", ",", "ctypes", ".", "c_char_p", ",", "ctypes", ".", "c_int", ",", "ctypes", ".", "c_int", ")", ")", "return", "f", "(", "p_mi", ",", "num", ",", "psz_filepath", ",", "i_width", ",", "i_height", ")" ]
take a snapshot of the current video window .
train
true
23,749
def unique_combinations(items, n): if (n == 0): (yield []) else: for i in xrange(((len(items) - n) + 1)): for cc in unique_combinations(items[(i + 1):], (n - 1)): (yield ([items[i]] + cc))
[ "def", "unique_combinations", "(", "items", ",", "n", ")", ":", "if", "(", "n", "==", "0", ")", ":", "(", "yield", "[", "]", ")", "else", ":", "for", "i", "in", "xrange", "(", "(", "(", "len", "(", "items", ")", "-", "n", ")", "+", "1", ")", ")", ":", "for", "cc", "in", "unique_combinations", "(", "items", "[", "(", "i", "+", "1", ")", ":", "]", ",", "(", "n", "-", "1", ")", ")", ":", "(", "yield", "(", "[", "items", "[", "i", "]", "]", "+", "cc", ")", ")" ]
return n-length tuples .
train
false
23,750
@Profiler.profile def test_dbapi_raw_w_connect(n): _test_dbapi_raw(n, True)
[ "@", "Profiler", ".", "profile", "def", "test_dbapi_raw_w_connect", "(", "n", ")", ":", "_test_dbapi_raw", "(", "n", ",", "True", ")" ]
individual insert/commit pairs w/ dbapi + connection each time .
train
false
23,751
def cr_uid(method): method._api = 'cr_uid' return method
[ "def", "cr_uid", "(", "method", ")", ":", "method", ".", "_api", "=", "'cr_uid'", "return", "method" ]
decorate a traditional-style method that takes cr .
train
false
23,754
def _convert_to_initializer(initializer): if isinstance(initializer, str): return getattr(tf, (initializer + '_initializer')) elif isinstance(initializer, np.ndarray): return tf.constant_initializer(initializer) else: return initializer
[ "def", "_convert_to_initializer", "(", "initializer", ")", ":", "if", "isinstance", "(", "initializer", ",", "str", ")", ":", "return", "getattr", "(", "tf", ",", "(", "initializer", "+", "'_initializer'", ")", ")", "elif", "isinstance", "(", "initializer", ",", "np", ".", "ndarray", ")", ":", "return", "tf", ".", "constant_initializer", "(", "initializer", ")", "else", ":", "return", "initializer" ]
returns a tensorflow initializer .
train
false
23,755
def search_paths_from_description(desc): paths = [] if desc.package: dirname = package_dirname(desc.package) paths.append(('', dirname)) elif desc.qualified_name: dirname = package_dirname(package(desc.qualified_name)) paths.append(('', dirname)) if hasattr(desc, 'search_paths'): paths.extend(desc.search_paths) return paths
[ "def", "search_paths_from_description", "(", "desc", ")", ":", "paths", "=", "[", "]", "if", "desc", ".", "package", ":", "dirname", "=", "package_dirname", "(", "desc", ".", "package", ")", "paths", ".", "append", "(", "(", "''", ",", "dirname", ")", ")", "elif", "desc", ".", "qualified_name", ":", "dirname", "=", "package_dirname", "(", "package", "(", "desc", ".", "qualified_name", ")", ")", "paths", ".", "append", "(", "(", "''", ",", "dirname", ")", ")", "if", "hasattr", "(", "desc", ",", "'search_paths'", ")", ":", "paths", ".", "extend", "(", "desc", ".", "search_paths", ")", "return", "paths" ]
return the search paths for the category/widgetdescription .
train
false
23,756
def auth_str_equal(provided, known): result = 0 p_len = len(provided) k_len = len(known) for i in xrange(p_len): a = (ord(provided[i]) if (i < p_len) else 0) b = (ord(known[i]) if (i < k_len) else 0) result |= (a ^ b) return ((p_len == k_len) & (result == 0))
[ "def", "auth_str_equal", "(", "provided", ",", "known", ")", ":", "result", "=", "0", "p_len", "=", "len", "(", "provided", ")", "k_len", "=", "len", "(", "known", ")", "for", "i", "in", "xrange", "(", "p_len", ")", ":", "a", "=", "(", "ord", "(", "provided", "[", "i", "]", ")", "if", "(", "i", "<", "p_len", ")", "else", "0", ")", "b", "=", "(", "ord", "(", "known", "[", "i", "]", ")", "if", "(", "i", "<", "k_len", ")", "else", "0", ")", "result", "|=", "(", "a", "^", "b", ")", "return", "(", "(", "p_len", "==", "k_len", ")", "&", "(", "result", "==", "0", ")", ")" ]
constant-time string comparison .
train
false
23,757
def demo_multiposition_feature(): postag(templates=[Template(Pos([(-3), (-2), (-1)]))])
[ "def", "demo_multiposition_feature", "(", ")", ":", "postag", "(", "templates", "=", "[", "Template", "(", "Pos", "(", "[", "(", "-", "3", ")", ",", "(", "-", "2", ")", ",", "(", "-", "1", ")", "]", ")", ")", "]", ")" ]
the feature/s of a template takes a list of positions relative to the current word where the feature should be looked for .
train
false
23,758
def _password_validators_help_text_html(password_validators=None): help_texts = password_validators_help_texts(password_validators) help_items = [format_html('<li>{}</li>', help_text) for help_text in help_texts] return (('<ul>%s</ul>' % ''.join(help_items)) if help_items else '')
[ "def", "_password_validators_help_text_html", "(", "password_validators", "=", "None", ")", ":", "help_texts", "=", "password_validators_help_texts", "(", "password_validators", ")", "help_items", "=", "[", "format_html", "(", "'<li>{}</li>'", ",", "help_text", ")", "for", "help_text", "in", "help_texts", "]", "return", "(", "(", "'<ul>%s</ul>'", "%", "''", ".", "join", "(", "help_items", ")", ")", "if", "help_items", "else", "''", ")" ]
return an html string with all help texts of all configured validators in an <ul> .
train
false
23,759
@utils.expects_func_args('instance') def reverts_task_state(function): @functools.wraps(function) def decorated_function(self, context, *args, **kwargs): try: return function(self, context, *args, **kwargs) except exception.UnexpectedTaskStateError as e: with excutils.save_and_reraise_exception(): LOG.info(_LI('Task possibly preempted: %s'), e.format_message()) except Exception: with excutils.save_and_reraise_exception(): wrapped_func = safe_utils.get_wrapped_function(function) keyed_args = inspect.getcallargs(wrapped_func, self, context, *args, **kwargs) instance = keyed_args['instance'] original_task_state = instance.task_state try: self._instance_update(context, instance, task_state=None) LOG.info(_LI('Successfully reverted task state from %s on failure for instance.'), original_task_state, instance=instance) except exception.InstanceNotFound: pass except Exception as e: msg = _LW('Failed to revert task state for instance. Error: %s') LOG.warning(msg, e, instance=instance) return decorated_function
[ "@", "utils", ".", "expects_func_args", "(", "'instance'", ")", "def", "reverts_task_state", "(", "function", ")", ":", "@", "functools", ".", "wraps", "(", "function", ")", "def", "decorated_function", "(", "self", ",", "context", ",", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "return", "function", "(", "self", ",", "context", ",", "*", "args", ",", "**", "kwargs", ")", "except", "exception", ".", "UnexpectedTaskStateError", "as", "e", ":", "with", "excutils", ".", "save_and_reraise_exception", "(", ")", ":", "LOG", ".", "info", "(", "_LI", "(", "'Task possibly preempted: %s'", ")", ",", "e", ".", "format_message", "(", ")", ")", "except", "Exception", ":", "with", "excutils", ".", "save_and_reraise_exception", "(", ")", ":", "wrapped_func", "=", "safe_utils", ".", "get_wrapped_function", "(", "function", ")", "keyed_args", "=", "inspect", ".", "getcallargs", "(", "wrapped_func", ",", "self", ",", "context", ",", "*", "args", ",", "**", "kwargs", ")", "instance", "=", "keyed_args", "[", "'instance'", "]", "original_task_state", "=", "instance", ".", "task_state", "try", ":", "self", ".", "_instance_update", "(", "context", ",", "instance", ",", "task_state", "=", "None", ")", "LOG", ".", "info", "(", "_LI", "(", "'Successfully reverted task state from %s on failure for instance.'", ")", ",", "original_task_state", ",", "instance", "=", "instance", ")", "except", "exception", ".", "InstanceNotFound", ":", "pass", "except", "Exception", "as", "e", ":", "msg", "=", "_LW", "(", "'Failed to revert task state for instance. Error: %s'", ")", "LOG", ".", "warning", "(", "msg", ",", "e", ",", "instance", "=", "instance", ")", "return", "decorated_function" ]
decorator to revert task_state on failure .
train
false
23,760
def validate_matrix_shape(name, shape, nrows, ncols, nobs): ndim = len(shape) if (ndim not in [2, 3]): raise ValueError(('Invalid value for %s matrix. Requires a 2- or 3-dimensional array, got %d dimensions' % (name, ndim))) if (not (shape[0] == nrows)): raise ValueError(('Invalid dimensions for %s matrix: requires %d rows, got %d' % (name, nrows, shape[0]))) if (not (shape[1] == ncols)): raise ValueError(('Invalid dimensions for %s matrix: requires %d columns, got %d' % (name, ncols, shape[1]))) if ((nobs is None) and (not ((ndim == 2) or (shape[(-1)] == 1)))): raise ValueError(('Invalid dimensions for %s matrix: time-varying matrices cannot be given unless `nobs` is specified (implicitly when a dataset is bound or else set explicity)' % name)) if ((ndim == 3) and (nobs is not None) and (not (shape[(-1)] in [1, nobs]))): raise ValueError(('Invalid dimensions for time-varying %s matrix. Requires shape (*,*,%d), got %s' % (name, nobs, str(shape))))
[ "def", "validate_matrix_shape", "(", "name", ",", "shape", ",", "nrows", ",", "ncols", ",", "nobs", ")", ":", "ndim", "=", "len", "(", "shape", ")", "if", "(", "ndim", "not", "in", "[", "2", ",", "3", "]", ")", ":", "raise", "ValueError", "(", "(", "'Invalid value for %s matrix. Requires a 2- or 3-dimensional array, got %d dimensions'", "%", "(", "name", ",", "ndim", ")", ")", ")", "if", "(", "not", "(", "shape", "[", "0", "]", "==", "nrows", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for %s matrix: requires %d rows, got %d'", "%", "(", "name", ",", "nrows", ",", "shape", "[", "0", "]", ")", ")", ")", "if", "(", "not", "(", "shape", "[", "1", "]", "==", "ncols", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for %s matrix: requires %d columns, got %d'", "%", "(", "name", ",", "ncols", ",", "shape", "[", "1", "]", ")", ")", ")", "if", "(", "(", "nobs", "is", "None", ")", "and", "(", "not", "(", "(", "ndim", "==", "2", ")", "or", "(", "shape", "[", "(", "-", "1", ")", "]", "==", "1", ")", ")", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for %s matrix: time-varying matrices cannot be given unless `nobs` is specified (implicitly when a dataset is bound or else set explicity)'", "%", "name", ")", ")", "if", "(", "(", "ndim", "==", "3", ")", "and", "(", "nobs", "is", "not", "None", ")", "and", "(", "not", "(", "shape", "[", "(", "-", "1", ")", "]", "in", "[", "1", ",", "nobs", "]", ")", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for time-varying %s matrix. Requires shape (*,*,%d), got %s'", "%", "(", "name", ",", "nobs", ",", "str", "(", "shape", ")", ")", ")", ")" ]
validate the shape of a possibly time-varying matrix .
train
false
23,761
def get_average(temp_base): if (not hasattr(get_average, 'temp')): get_average.temp = [temp_base, temp_base, temp_base] get_average.temp[2] = get_average.temp[1] get_average.temp[1] = get_average.temp[0] get_average.temp[0] = temp_base temp_avg = (((get_average.temp[0] + get_average.temp[1]) + get_average.temp[2]) / 3) return temp_avg
[ "def", "get_average", "(", "temp_base", ")", ":", "if", "(", "not", "hasattr", "(", "get_average", ",", "'temp'", ")", ")", ":", "get_average", ".", "temp", "=", "[", "temp_base", ",", "temp_base", ",", "temp_base", "]", "get_average", ".", "temp", "[", "2", "]", "=", "get_average", ".", "temp", "[", "1", "]", "get_average", ".", "temp", "[", "1", "]", "=", "get_average", ".", "temp", "[", "0", "]", "get_average", ".", "temp", "[", "0", "]", "=", "temp_base", "temp_avg", "=", "(", "(", "(", "get_average", ".", "temp", "[", "0", "]", "+", "get_average", ".", "temp", "[", "1", "]", ")", "+", "get_average", ".", "temp", "[", "2", "]", ")", "/", "3", ")", "return", "temp_avg" ]
use moving average to get better readings .
train
false
23,763
@profiler.trace def rule_list_for_tenant(request, tenant_id, **kwargs): rules = rule_list(request, tenant_id=tenant_id, shared=False, **kwargs) shared_rules = rule_list(request, shared=True, **kwargs) return (rules + shared_rules)
[ "@", "profiler", ".", "trace", "def", "rule_list_for_tenant", "(", "request", ",", "tenant_id", ",", "**", "kwargs", ")", ":", "rules", "=", "rule_list", "(", "request", ",", "tenant_id", "=", "tenant_id", ",", "shared", "=", "False", ",", "**", "kwargs", ")", "shared_rules", "=", "rule_list", "(", "request", ",", "shared", "=", "True", ",", "**", "kwargs", ")", "return", "(", "rules", "+", "shared_rules", ")" ]
return a rule list available for the tenant .
train
false
23,765
def _future_expose_api_anonymous_and_sessionless(func, to_json=True): return _future_expose_api(func, to_json=to_json, user_required=False, user_or_session_required=False)
[ "def", "_future_expose_api_anonymous_and_sessionless", "(", "func", ",", "to_json", "=", "True", ")", ":", "return", "_future_expose_api", "(", "func", ",", "to_json", "=", "to_json", ",", "user_required", "=", "False", ",", "user_or_session_required", "=", "False", ")" ]
expose this function via the api but dont require a user or a galaxy_session .
train
false
23,766
def running_under_virtualenv(): return hasattr(sys, 'real_prefix')
[ "def", "running_under_virtualenv", "(", ")", ":", "return", "hasattr", "(", "sys", ",", "'real_prefix'", ")" ]
return true if were running inside a virtualenv .
train
false
23,767
def setitem(a, b, c): a[b] = c
[ "def", "setitem", "(", "a", ",", "b", ",", "c", ")", ":", "a", "[", "b", "]", "=", "c" ]
same as a[b] = c .
train
false
23,768
def show_offload(devname): try: sg = ((ethtool.get_sg(devname) and 'on') or 'off') except IOError: sg = 'not supported' try: tso = ((ethtool.get_tso(devname) and 'on') or 'off') except IOError: tso = 'not supported' try: ufo = ((ethtool.get_ufo(devname) and 'on') or 'off') except IOError: ufo = 'not supported' try: gso = ((ethtool.get_gso(devname) and 'on') or 'off') except IOError: gso = 'not supported' offload = {'scatter_gather': sg, 'tcp_segmentation_offload': tso, 'udp_fragmentation_offload': ufo, 'generic_segmentation_offload': gso} return offload
[ "def", "show_offload", "(", "devname", ")", ":", "try", ":", "sg", "=", "(", "(", "ethtool", ".", "get_sg", "(", "devname", ")", "and", "'on'", ")", "or", "'off'", ")", "except", "IOError", ":", "sg", "=", "'not supported'", "try", ":", "tso", "=", "(", "(", "ethtool", ".", "get_tso", "(", "devname", ")", "and", "'on'", ")", "or", "'off'", ")", "except", "IOError", ":", "tso", "=", "'not supported'", "try", ":", "ufo", "=", "(", "(", "ethtool", ".", "get_ufo", "(", "devname", ")", "and", "'on'", ")", "or", "'off'", ")", "except", "IOError", ":", "ufo", "=", "'not supported'", "try", ":", "gso", "=", "(", "(", "ethtool", ".", "get_gso", "(", "devname", ")", "and", "'on'", ")", "or", "'off'", ")", "except", "IOError", ":", "gso", "=", "'not supported'", "offload", "=", "{", "'scatter_gather'", ":", "sg", ",", "'tcp_segmentation_offload'", ":", "tso", ",", "'udp_fragmentation_offload'", ":", "ufo", ",", "'generic_segmentation_offload'", ":", "gso", "}", "return", "offload" ]
queries the specified network device for the state of protocol offload and other features cli example: .
train
true