idx int64 0 63k | question stringlengths 61 4.03k | target stringlengths 6 1.23k |
|---|---|---|
51,000 | def dump ( self , fields = None , exclude = None ) : exclude = exclude or [ ] d = { } if fields and self . _primary_field not in fields : fields = list ( fields ) fields . append ( self . _primary_field ) for k , v in self . properties . items ( ) : if ( ( not fields ) or ( k in fields ) ) and ( not exclude or ( k not in exclude ) ) : if not isinstance ( v , ManyToMany ) : t = v . get_value_for_datastore ( self ) if t is Lazy : self . refresh ( ) t = v . get_value_for_datastore ( self ) if isinstance ( t , Model ) : t = t . _key d [ k ] = v . to_str ( t ) else : if fields : d [ k ] = ',' . join ( [ str ( x ) for x in getattr ( self , v . _lazy_value ( ) , [ ] ) ] ) if self . _primary_field and d and self . _primary_field not in d : d [ self . _primary_field ] = str ( self . _key ) return d | Dump current object to dict but the value is string for manytomany fields will not automatically be dumpped only when they are given in fields parameter |
51,001 | def clear_relation ( cls ) : for k , v in cls . properties . items ( ) : if isinstance ( v , ReferenceProperty ) : if hasattr ( v , 'collection_name' ) and hasattr ( v . reference_class , v . collection_name ) : delattr ( v . reference_class , v . collection_name ) if isinstance ( v , OneToOne ) : del v . reference_class . _onetoone [ v . collection_name ] | Clear relation properties for reference Model such as OneToOne Reference ManyToMany |
51,002 | def put ( self , _name , ** values ) : try : sql = self . sqles [ _name ] data = sql [ 'data' ] if sql [ 'positional' ] : d = [ values [ k ] for k , v in sql [ 'fields' ] . items ( ) ] else : d = { v : values [ k ] for k , v in sql [ 'fields' ] . items ( ) } data . append ( d ) if self . size and len ( data ) >= self . size : do_ ( sql [ 'raw_sql' ] , args = data ) sql [ 'data' ] = [ ] except : if self . transcation : Rollback ( self . engine ) raise | Put data to cach if reached size value it ll execute at once . |
51,003 | def get_key ( keyfile = None ) : keyfile = keyfile or application_path ( settings . SECRETKEY . SECRET_FILE ) with file ( keyfile , 'rb' ) as f : return f . read ( ) | Read the key content from secret_file |
51,004 | def get_cipher_key ( keyfile = None ) : _key = get_key ( keyfile ) _k = md5 ( _key ) . hexdigest ( ) key = xor ( _k [ : 8 ] , _k [ 8 : 16 ] , _k [ 16 : 24 ] , _k [ 24 : ] ) return key | Create key which will be used in des because des need 8bytes chars |
51,005 | def set_many ( self , mapping , timeout = None ) : for key , value in _items ( mapping ) : self . set ( key , value , timeout ) | Sets multiple keys and values from a mapping . |
51,006 | def dec ( self , key , delta = 1 ) : self . set ( key , ( self . get ( key ) or 0 ) - delta ) | Decrements the value of a key by delta . If the key does not yet exist it is initialized with - delta . |
51,007 | def import_preferred_memcache_lib ( self , servers ) : try : import pylibmc except ImportError : pass else : return pylibmc . Client ( servers ) try : from google . appengine . api import memcache except ImportError : pass else : return memcache . Client ( ) try : import memcache except ImportError : pass else : return memcache . Client ( servers ) | Returns an initialized memcache client . Used by the constructor . |
51,008 | def dump_object ( self , value ) : t = type ( value ) if t is int or t is long : return str ( value ) return '!' + pickle . dumps ( value ) | Dumps an object into a string for redis . By default it serializes integers as regular string and pickle dumps everything else . |
51,009 | def extract_dirs ( mod , path , dst , verbose = False , exclude = None , exclude_ext = None , recursion = True , replace = True ) : default_exclude = [ '.svn' , '_svn' , '.git' ] default_exclude_ext = [ '.pyc' , '.pyo' , '.bak' , '.tmp' ] exclude = exclude or [ ] exclude_ext = exclude_ext or [ ] if not os . path . exists ( dst ) : os . makedirs ( dst ) if verbose : print 'Make directory %s' % dst for r in pkg . resource_listdir ( mod , path ) : if r in exclude or r in default_exclude : continue fpath = os . path . join ( path , r ) if pkg . resource_isdir ( mod , fpath ) : if recursion : extract_dirs ( mod , fpath , os . path . join ( dst , r ) , verbose , exclude , exclude_ext , recursion , replace ) else : ext = os . path . splitext ( fpath ) [ 1 ] if ext in exclude_ext or ext in default_exclude_ext : continue extract_file ( mod , fpath , dst , verbose , replace ) | mod name path mod path dst output directory resursion True will extract all sub module of mod |
51,010 | def walk_dirs ( path , include = None , include_ext = None , exclude = None , exclude_ext = None , recursion = True , file_only = False , use_default_pattern = True , patterns = None ) : default_exclude = [ '.svn' , '_svn' , '.git' ] default_exclude_ext = [ '.pyc' , '.pyo' , '.bak' , '.tmp' ] exclude = exclude or [ ] exclude_ext = exclude_ext or [ ] include_ext = include_ext or [ ] include = include or [ ] if not os . path . exists ( path ) : raise StopIteration for r in os . listdir ( path ) : if match ( r , exclude ) or ( use_default_pattern and r in default_exclude ) : continue if include and r not in include : continue fpath = os . path . join ( path , r ) if os . path . isdir ( fpath ) : if not file_only : if patterns and match ( r , patterns ) : yield os . path . normpath ( fpath ) . replace ( '\\' , '/' ) if recursion : for f in walk_dirs ( fpath , include , include_ext , exclude , exclude_ext , recursion , file_only , use_default_pattern , patterns ) : yield os . path . normpath ( f ) . replace ( '\\' , '/' ) else : ext = os . path . splitext ( fpath ) [ 1 ] if ext in exclude_ext or ( use_default_pattern and ext in default_exclude_ext ) : continue if include_ext and ext not in include_ext : continue if patterns : if not match ( r , patterns ) : continue yield os . path . normpath ( fpath ) . replace ( '\\' , '/' ) | path directory path resursion True will extract all sub module of mod |
51,011 | def camel_to_ ( s ) : s1 = re . sub ( '(.)([A-Z][a-z]+)' , r'\1_\2' , s ) return re . sub ( '([a-z0-9])([A-Z])' , r'\1_\2' , s1 ) . lower ( ) | Convert CamelCase to camel_case |
51,012 | def application_path ( path ) : from uliweb import application return os . path . join ( application . project_dir , path ) | Join application project_dir and path |
51,013 | def get_uuid ( type = 4 ) : import uuid name = 'uuid' + str ( type ) u = getattr ( uuid , name ) return u ( ) . hex | Get uuid value |
51,014 | def request_url ( req = None ) : from uliweb import request r = req or request if request : if r . query_string : return r . path + '?' + r . query_string else : return r . path else : return '' | Get full url of a request |
51,015 | def compare_dict ( da , db ) : sa = set ( da . items ( ) ) sb = set ( db . items ( ) ) diff = sa & sb return dict ( sa - diff ) , dict ( sb - diff ) | Compare differencs from two dicts |
51,016 | def get_configrable_object ( key , section , cls = None ) : from uliweb import UliwebError , settings import inspect if inspect . isclass ( key ) and cls and issubclass ( key , cls ) : return key elif isinstance ( key , ( str , unicode ) ) : path = settings [ section ] . get ( key ) if path : _cls = import_attr ( path ) return _cls else : raise UliwebError ( "Can't find section name %s in settings" % section ) else : raise UliwebError ( "Key %r should be subclass of %r object or string path format!" % ( key , cls ) ) | if obj is a class then check if the class is subclass of cls or it should be object path and it ll be imported by import_attr |
51,017 | def convert_bytes ( n ) : symbols = ( 'K' , 'M' , 'G' , 'T' , 'P' , 'E' , 'Z' , 'Y' ) prefix = { } for i , s in enumerate ( symbols ) : prefix [ s ] = 1 << ( i + 1 ) * 10 for s in reversed ( symbols ) : if n >= prefix [ s ] : value = float ( n ) / prefix [ s ] return '%.1f%s' % ( value , s ) return "%sB" % n | Convert a size number to K M . etc |
51,018 | def init_static_combine ( ) : from uliweb import settings from hashlib import md5 import os d = { } if settings . get_var ( 'STATIC_COMBINE_CONFIG/enable' , False ) : for k , v in settings . get ( 'STATIC_COMBINE' , { } ) . items ( ) : key = '_cmb_' + md5 ( '' . join ( v ) ) . hexdigest ( ) + os . path . splitext ( v [ 0 ] ) [ 1 ] d [ key ] = v return d | Process static combine create md5 key according each static filename |
51,019 | def csrf_token ( ) : from uliweb import request , settings from uliweb . utils . common import safe_str v = { } token_name = settings . CSRF . cookie_token_name if not request . session . deleted and request . session . get ( token_name ) : v = request . session [ token_name ] if time . time ( ) >= v [ 'created_time' ] + v [ 'expiry_time' ] : v = { } else : v [ 'created_time' ] = time . time ( ) if not v : token = request . cookies . get ( token_name ) if not token : token = uuid . uuid4 ( ) . get_hex ( ) v = { 'token' : token , 'expiry_time' : settings . CSRF . timeout , 'created_time' : time . time ( ) } if not request . session . deleted : request . session [ token_name ] = v return safe_str ( v [ 'token' ] ) | Get csrf token or create new one |
51,020 | def json ( self ) : if 'json' not in self . environ . get ( 'CONTENT_TYPE' , '' ) : raise BadRequest ( 'Not a JSON request' ) try : return loads ( self . data ) except Exception : raise BadRequest ( 'Unable to read JSON request' ) | Get the result of simplejson . loads if possible . |
51,021 | def parse_protobuf ( self , proto_type ) : if 'protobuf' not in self . environ . get ( 'CONTENT_TYPE' , '' ) : raise BadRequest ( 'Not a Protobuf request' ) obj = proto_type ( ) try : obj . ParseFromString ( self . data ) except Exception : raise BadRequest ( "Unable to parse Protobuf request" ) if self . protobuf_check_initialization and not obj . IsInitialized ( ) : raise BadRequest ( "Partial Protobuf request" ) return obj | Parse the data into an instance of proto_type . |
51,022 | def charset ( self ) : header = self . environ . get ( 'CONTENT_TYPE' ) if header : ct , options = parse_options_header ( header ) charset = options . get ( 'charset' ) if charset : if is_known_charset ( charset ) : return charset return self . unknown_charset ( charset ) return self . default_charset | The charset from the content type . |
51,023 | def utf8 ( value ) : if isinstance ( value , _UTF8_TYPES ) : return value elif isinstance ( value , unicode_type ) : return value . encode ( "utf-8" ) else : return str ( value ) | Converts a string argument to a byte string . If the argument is already a byte string or None it is returned unchanged . Otherwise it must be a unicode string and is encoded as utf8 . |
51,024 | def to_basestring ( value ) : if value is None : return 'None' if isinstance ( value , _BASESTRING_TYPES ) : return value elif isinstance ( value , unicode_type ) : return value . decode ( "utf-8" ) else : return str ( value ) | Converts a string argument to a subclass of basestring . In python2 byte and unicode strings are mostly interchangeable so functions that deal with a user - supplied argument in combination with ascii string constants can use either and should return the type the user supplied . In python3 the two types are not interchangeable so this method is needed to convert byte strings to unicode . |
51,025 | def clear ( self ) : self . __values . clear ( ) self . __access_keys = [ ] self . __modified_times . clear ( ) | Clears the dict . |
51,026 | def has ( self , key , mtime = None ) : v = self . __values . get ( key , None ) if not v : return False if self . check_modified_time : mtime = self . _get_mtime ( key , mtime ) if mtime != self . __modified_times [ key ] : del self [ key ] return False return True | This method should almost NEVER be used . The reason is that between the time has_key is called and the key is accessed the key might vanish . |
51,027 | def reset ( self ) : with self . lock : if self . cache : if self . use_tmp : shutil . rmtree ( self . tmp_dir , ignore_errors = True ) else : self . templates = { } | Resets the cache of compiled templates . |
51,028 | def has_role ( user , * roles , ** kwargs ) : Role = get_model ( 'role' ) if isinstance ( user , ( unicode , str ) ) : User = get_model ( 'user' ) user = User . get ( User . c . username == user ) for role in roles : if isinstance ( role , ( str , unicode ) ) : role = Role . get ( Role . c . name == role ) if not role : continue name = role . name func = __role_funcs__ . get ( name , None ) if func : if isinstance ( func , ( unicode , str ) ) : func = import_attr ( func ) assert callable ( func ) para = kwargs . copy ( ) para [ 'user' ] = user flag = call_func ( func , para ) if flag : return role flag = role . users . has ( user ) if flag : return role flag = role . usergroups_has_user ( user ) if flag : return role return False | Judge is the user belongs to the role and if does then return the role object if not then return False . kwargs will be passed to role_func . |
51,029 | def has_permission ( user , * permissions , ** role_kwargs ) : Role = get_model ( 'role' ) Perm = get_model ( 'permission' ) if isinstance ( user , ( unicode , str ) ) : User = get_model ( 'user' ) user = User . get ( User . c . username == user ) for name in permissions : perm = Perm . get ( Perm . c . name == name ) if not perm : continue flag = has_role ( user , * list ( perm . perm_roles . with_relation ( ) . all ( ) ) , ** role_kwargs ) if flag : return flag return False | Judge if an user has permission and if it does return role object and if it doesn t return False . role_kwargs will be passed to role functions . With role object you can use role . relation to get Role_Perm_Rel object . |
51,030 | def load_ssl_context ( cert_file , pkey_file ) : from OpenSSL import SSL ctx = SSL . Context ( SSL . SSLv23_METHOD ) ctx . use_certificate_file ( cert_file ) ctx . use_privatekey_file ( pkey_file ) return ctx | Loads an SSL context from a certificate and private key file . |
51,031 | def select_ip_version ( host , port ) : if ':' in host and hasattr ( socket , 'AF_INET6' ) : return socket . AF_INET6 return socket . AF_INET | Returns AF_INET4 or AF_INET6 depending on where to connect to . |
51,032 | def make_server ( host , port , app = None , threaded = False , processes = 1 , request_handler = None , passthrough_errors = False , ssl_context = None ) : if threaded and processes > 1 : raise ValueError ( "cannot have a multithreaded and " "multi process server." ) elif threaded : return ThreadedWSGIServer ( host , port , app , request_handler , passthrough_errors , ssl_context ) elif processes > 1 : return ForkingWSGIServer ( host , port , app , processes , request_handler , passthrough_errors , ssl_context ) else : return BaseWSGIServer ( host , port , app , request_handler , passthrough_errors , ssl_context ) | Create a new server instance that is either threaded or forks or just processes one request after another . |
51,033 | def _reloader_stat_loop ( extra_files = None , interval = 1 ) : from itertools import chain mtimes = { } while 1 : for filename in chain ( _iter_module_files ( ) , extra_files or ( ) ) : try : mtime = os . stat ( filename ) . st_mtime except OSError : continue old_time = mtimes . get ( filename ) if old_time is None : mtimes [ filename ] = mtime continue elif mtime > old_time : _log ( 'info' , ' * Detected change in %r, reloading' % filename ) sys . exit ( 3 ) time . sleep ( interval ) | When this function is run from the main thread it will force other threads to exit when any modules currently loaded change . |
51,034 | def run_simple ( hostname , port , application , use_reloader = False , use_debugger = False , use_evalex = True , extra_files = None , reloader_interval = 1 , threaded = False , processes = 1 , request_handler = None , static_files = None , passthrough_errors = False , ssl_context = None ) : if use_debugger : from werkzeug . debug import DebuggedApplication application = DebuggedApplication ( application , use_evalex ) if static_files : from werkzeug . wsgi import SharedDataMiddleware application = SharedDataMiddleware ( application , static_files ) def inner ( ) : make_server ( hostname , port , application , threaded , processes , request_handler , passthrough_errors , ssl_context ) . serve_forever ( ) if os . environ . get ( 'WERKZEUG_RUN_MAIN' ) != 'true' : display_hostname = hostname != '*' and hostname or 'localhost' if ':' in display_hostname : display_hostname = '[%s]' % display_hostname _log ( 'info' , ' * Running on %s://%s:%d/' , ssl_context is None and 'http' or 'https' , display_hostname , port ) if use_reloader : address_family = select_ip_version ( hostname , port ) test_socket = socket . socket ( address_family , socket . SOCK_STREAM ) test_socket . setsockopt ( socket . SOL_SOCKET , socket . SO_REUSEADDR , 1 ) test_socket . bind ( ( hostname , port ) ) test_socket . close ( ) run_with_reloader ( inner , extra_files , reloader_interval ) else : inner ( ) | Start an application using wsgiref and with an optional reloader . This wraps wsgiref to fix the wrong default reporting of the multithreaded WSGI variable and adds optional multithreading and fork support . |
51,035 | def to_attrs ( args , nocreate_if_none = [ 'id' , 'for' , 'class' ] ) : if not args : return '' s = [ '' ] for k , v in sorted ( args . items ( ) ) : k = u_str ( k ) v = u_str ( v ) if k . startswith ( '_' ) : k = k [ 1 : ] if v is None : if k not in nocreate_if_none : s . append ( k ) else : if k . lower ( ) in __noescape_attrs__ : t = u_str ( v ) else : t = cgi . escape ( u_str ( v ) ) t = '"%s"' % t . replace ( '"' , '"' ) s . append ( '%s=%s' % ( k , t ) ) return ' ' . join ( s ) | Make python dict to k = v format |
51,036 | def _get_modpkg_path ( dotted_name , pathlist = None ) : parts = dotted_name . split ( '.' , 1 ) if len ( parts ) > 1 : try : file , pathname , description = imp . find_module ( parts [ 0 ] , pathlist ) if file : file . close ( ) except ImportError : return None if description [ 2 ] == imp . PKG_DIRECTORY : pathname = _get_modpkg_path ( parts [ 1 ] , [ pathname ] ) else : pathname = None else : try : file , pathname , description = imp . find_module ( dotted_name , pathlist ) if file : file . close ( ) if description [ 2 ] not in [ imp . PY_SOURCE , imp . PKG_DIRECTORY ] : pathname = None except ImportError : pathname = None return pathname | Get the filesystem path for a module or a package . |
51,037 | def getFilesForName ( name ) : if not os . path . exists ( name ) : if containsAny ( name , "*?[]" ) : files = glob . glob ( name ) list = [ ] for file in files : list . extend ( getFilesForName ( file ) ) return list name = _get_modpkg_path ( name ) if not name : return [ ] if os . path . isdir ( name ) : list = [ ] os . path . walk ( name , _visit_pyfiles , list ) return list elif os . path . exists ( name ) : return [ name ] return [ ] | Get a list of module files for a filename a module or package name or a directory . |
51,038 | def unescape ( s ) : def handle_match ( m ) : name = m . group ( 1 ) if name in HTMLBuilder . _entities : return unichr ( HTMLBuilder . _entities [ name ] ) try : if name [ : 2 ] in ( '#x' , '#X' ) : return unichr ( int ( name [ 2 : ] , 16 ) ) elif name . startswith ( '#' ) : return unichr ( int ( name [ 1 : ] ) ) except ValueError : pass return u'' return _entity_re . sub ( handle_match , s ) | The reverse function of escape . This unescapes all the HTML entities not only the XML entities inserted by escape . |
51,039 | def append_slash_redirect ( environ , code = 301 ) : new_path = environ [ 'PATH_INFO' ] . strip ( '/' ) + '/' query_string = environ . get ( 'QUERY_STRING' ) if query_string : new_path += '?' + query_string return redirect ( new_path , code ) | Redirect to the same URL but with a slash appended . The behavior of this function is undefined if the path ends with a slash already . |
51,040 | def parse_translation ( f , lineno ) : line = f . readline ( ) def get_line ( f , line , need_keys , lineno , default = '""' ) : line = line . rstrip ( ) if not line : return lineno , need_keys [ 0 ] , default , line key , value = line . split ( ' ' , 1 ) if key not in need_keys : print 'Error Line, need %r: %d, line=' % ( need_keys , lineno , line ) raise RuntimeError ( "parse error" ) v = value while 1 : line = f . readline ( ) line = line . rstrip ( ) lineno += 1 if not line or line [ 0 ] != '"' : break v += '\n' + line [ : ] return lineno , key , v , line comments = [ ] while 1 : if not line : return lineno , None , None , None if line . strip ( ) == '' : return lineno , comments , None , None elif line [ 0 ] == '#' : comments . append ( line [ : - 1 ] ) else : break line = f . readline ( ) lineno += 1 lineno , key , msgid , line = get_line ( f , line , [ 'msgid' ] , lineno ) lineno , key , value , line = get_line ( f , line , [ 'msgid_plural' , 'msgstr' ] , lineno ) if key == 'msgid_plural' : msgid = ( msgid , value ) lineno , key , v1 , line = get_line ( f , line , [ 'msgstr[0]' ] , lineno ) lineno , key , v2 , line = get_line ( f , line , [ 'msgstr[1]' ] , lineno ) msgstr = ( v1 , v2 ) else : msgstr = value if line != '' : print 'File: %s Error Line: %s' % ( f . name , line ) raise RuntimeError ( "parse error" ) return lineno , comments , msgid , msgstr | Read a single translation entry from the file F and return a tuple with the comments msgid and msgstr . The comments is returned as a list of lines which do not end in new - lines . The msgid and msgstr are strings possibly with embedded newlines |
51,041 | def get_form ( formcls ) : from uliweb . form import Form import inspect if inspect . isclass ( formcls ) and issubclass ( formcls , Form ) : return formcls elif isinstance ( formcls , ( str , unicode ) ) : path = settings . FORMS . get ( formcls ) if path : _cls = import_attr ( path ) return _cls else : raise UliwebError ( "Can't find formcls name %s in settings.FORMS" % formcls ) else : raise UliwebError ( "formcls should be Form class object or string path format, but %r found!" % formcls ) | get form class according form class path or form class object |
51,042 | def run ( namespace = None , action_prefix = 'action_' , args = None ) : if namespace is None : namespace = sys . _getframe ( 1 ) . f_locals actions = find_actions ( namespace , action_prefix ) if args is None : args = sys . argv [ 1 : ] if not args or args [ 0 ] in ( '-h' , '--help' ) : return print_usage ( actions ) elif args [ 0 ] not in actions : fail ( 'Unknown action \'%s\'' % args [ 0 ] ) arguments = { } types = { } key_to_arg = { } long_options = [ ] formatstring = '' func , doc , arg_def = actions [ args . pop ( 0 ) ] for idx , ( arg , shortcut , default , option_type ) in enumerate ( arg_def ) : real_arg = arg . replace ( '-' , '_' ) if shortcut : formatstring += shortcut if not isinstance ( default , bool ) : formatstring += ':' key_to_arg [ '-' + shortcut ] = real_arg long_options . append ( isinstance ( default , bool ) and arg or arg + '=' ) key_to_arg [ '--' + arg ] = real_arg key_to_arg [ idx ] = real_arg types [ real_arg ] = option_type arguments [ real_arg ] = default try : optlist , posargs = getopt . gnu_getopt ( args , formatstring , long_options ) except getopt . GetoptError as e : fail ( str ( e ) ) specified_arguments = set ( ) for key , value in enumerate ( posargs ) : try : arg = key_to_arg [ key ] except IndexError : fail ( 'Too many parameters' ) specified_arguments . add ( arg ) try : arguments [ arg ] = converters [ types [ arg ] ] ( value ) except ValueError : fail ( 'Invalid value for argument %s (%s): %s' % ( key , arg , value ) ) for key , value in optlist : arg = key_to_arg [ key ] if arg in specified_arguments : fail ( 'Argument \'%s\' is specified twice' % arg ) if types [ arg ] == 'boolean' : if arg . startswith ( 'no_' ) : value = 'no' else : value = 'yes' try : arguments [ arg ] = converters [ types [ arg ] ] ( value ) except ValueError : fail ( 'Invalid value for \'%s\': %s' % ( key , value ) ) newargs = { } for k , v in iteritems ( arguments ) : newargs [ k . startswith ( 'no_' ) and k [ 3 : ] or k ] = v arguments = newargs return func ( ** arguments ) | Run the script . Participating actions are looked up in the caller s namespace if no namespace is given otherwise in the dict provided . Only items that start with action_prefix are processed as actions . If you want to use all items in the namespace provided as actions set action_prefix to an empty string . |
51,043 | def fail ( message , code = - 1 ) : print ( 'Error: %s' % message , file = sys . stderr ) sys . exit ( code ) | Fail with an error . |
51,044 | def find_actions ( namespace , action_prefix ) : actions = { } for key , value in iteritems ( namespace ) : if key . startswith ( action_prefix ) : actions [ key [ len ( action_prefix ) : ] ] = analyse_action ( value ) return actions | Find all the actions in the namespace . |
51,045 | def analyse_action ( func ) : description = inspect . getdoc ( func ) or 'undocumented action' arguments = [ ] args , varargs , kwargs , defaults = inspect . getargspec ( func ) if varargs or kwargs : raise TypeError ( 'variable length arguments for action not allowed.' ) if len ( args ) != len ( defaults or ( ) ) : raise TypeError ( 'not all arguments have proper definitions' ) for idx , ( arg , definition ) in enumerate ( zip ( args , defaults or ( ) ) ) : if arg . startswith ( '_' ) : raise TypeError ( 'arguments may not start with an underscore' ) if not isinstance ( definition , tuple ) : shortcut = None default = definition else : shortcut , default = definition argument_type = argument_types [ type ( default ) ] if isinstance ( default , bool ) and default is True : arg = 'no-' + arg arguments . append ( ( arg . replace ( '_' , '-' ) , shortcut , default , argument_type ) ) return func , description , arguments | Analyse a function . |
51,046 | def make_shell ( init_func = None , banner = None , use_ipython = True ) : if banner is None : banner = 'Interactive Werkzeug Shell' if init_func is None : init_func = dict def action ( ipython = use_ipython ) : namespace = init_func ( ) if ipython : try : try : from IPython . frontend . terminal . embed import InteractiveShellEmbed sh = InteractiveShellEmbed ( banner1 = banner ) except ImportError : from IPython . Shell import IPShellEmbed sh = IPShellEmbed ( banner = banner ) except ImportError : pass else : sh ( global_ns = { } , local_ns = namespace ) return from code import interact interact ( banner , local = namespace ) return action | Returns an action callback that spawns a new interactive python shell . |
51,047 | def make_runserver ( app_factory , hostname = 'localhost' , port = 5000 , use_reloader = False , use_debugger = False , use_evalex = True , threaded = False , processes = 1 , static_files = None , extra_files = None , ssl_context = None ) : def action ( hostname = ( 'h' , hostname ) , port = ( 'p' , port ) , reloader = use_reloader , debugger = use_debugger , evalex = use_evalex , threaded = threaded , processes = processes ) : from werkzeug . serving import run_simple app = app_factory ( ) run_simple ( hostname , port , app , reloader , debugger , evalex , extra_files , 1 , threaded , processes , static_files = static_files , ssl_context = ssl_context ) return action | Returns an action callback that spawns a new development server . |
51,048 | def unbind ( topic , func ) : if topic in _receivers : receivers = _receivers [ topic ] for i in range ( len ( receivers ) - 1 , - 1 , - 1 ) : nice , f = receivers [ i ] if ( callable ( func ) and f [ 'func' ] == func ) or ( f [ 'func_name' ] == func ) : del receivers [ i ] return | Remove receiver function |
51,049 | def call ( sender , topic , * args , ** kwargs ) : if not topic in _receivers : return items = _receivers [ topic ] def _cmp ( x , y ) : return cmp ( x [ 0 ] , y [ 0 ] ) items . sort ( _cmp ) i = 0 while i < len ( items ) : nice , f = items [ i ] i = i + 1 _f = f [ 'func' ] if not _f : try : _f = import_attr ( f [ 'func_name' ] ) except ( ImportError , AttributeError ) as e : logging . error ( "Can't import function %s" % f [ 'func_name' ] ) raise f [ 'func' ] = _f if callable ( _f ) : kw = kwargs . copy ( ) if not _test ( kw , f ) : continue try : _f ( sender , * args , ** kw ) except : func = _f . __module__ + '.' + _f . __name__ logging . exception ( 'Calling dispatch point [%s] %s(%r, %r) error!' % ( topic , func , args , kw ) ) raise else : raise Exception , "Dispatch point [%s] %r can't been invoked" % ( topic , _f ) | Invoke receiver functions according topic it ll invoke receiver functions one by one and it ll not return anything so if you want to return a value you should use get function . |
51,050 | def get ( sender , topic , * args , ** kwargs ) : if not topic in _receivers : return items = _receivers [ topic ] def _cmp ( x , y ) : return cmp ( x [ 0 ] , y [ 0 ] ) items . sort ( _cmp ) for i in range ( len ( items ) ) : nice , f = items [ i ] _f = f [ 'func' ] if not _f : try : _f = import_attr ( f [ 'func_name' ] ) except ImportError : logging . error ( "Can't import function %s" % f [ 'func_name' ] ) raise f [ 'func' ] = _f if callable ( _f ) : if not _test ( kwargs , f ) : continue try : v = _f ( sender , * args , ** kwargs ) except : func = _f . __module__ + '.' + _f . __name__ logging . exception ( 'Calling dispatch point [%s] %s(%r,%r) error!' % ( topic , func , args , kwargs ) ) raise if v is not None : return v else : raise "Dispatch point [%s] %r can't been invoked" % ( topic , _f ) | Invoke receiver functions according topic it ll invoke receiver functions one by one and if one receiver function return non - None value it ll return it and break the loop . |
51,051 | def __read_line ( self , f ) : g = tokenize . generate_tokens ( f . readline ) buf = [ ] time = 0 iden_existed = False while 1 : v = g . next ( ) tokentype , t , start , end , line = v if tokentype == 54 : continue if tokentype in ( token . INDENT , token . DEDENT , tokenize . COMMENT ) : continue if tokentype == token . NAME : iden_existed = True if tokentype == token . NEWLINE : return '' . join ( buf ) , iden_existed else : if t == '=' and time == 0 : time += 1 continue buf . append ( t ) | Get logic line according the syntax not the physical line It ll return the line text and if there is identifier existed return line bool |
51,052 | def freeze ( self ) : self . _lazy = False for k , v in self . items ( ) : if k in self . _env : continue for _k , _v in v . items ( ) : if isinstance ( _v , Lazy ) : if self . writable : _v . get ( ) else : try : v . __setitem__ ( _k , _v . get ( ) , replace = True ) except : print "Error ini key:" , _k raise del _v self . _globals = SortedDict ( ) | Process all EvalValue to real value |
51,053 | def serialize ( self , expires = None ) : if self . secret_key is None : raise RuntimeError ( 'no secret key defined' ) if expires : self [ '_expires' ] = _date_to_unix ( expires ) result = [ ] mac = hmac ( self . secret_key , None , self . hash_method ) for key , value in sorted ( self . items ( ) ) : result . append ( ( '%s=%s' % ( url_quote_plus ( key ) , self . quote ( value ) . decode ( 'ascii' ) ) ) . encode ( 'ascii' ) ) mac . update ( b'|' + result [ - 1 ] ) return b'?' . join ( [ base64 . b64encode ( mac . digest ( ) ) . strip ( ) , b'&' . join ( result ) ] ) | Serialize the secure cookie into a string . |
51,054 | def unserialize ( cls , string , secret_key ) : if isinstance ( string , text_type ) : string = string . encode ( 'utf-8' , 'replace' ) if isinstance ( secret_key , text_type ) : secret_key = secret_key . encode ( 'utf-8' , 'replace' ) try : base64_hash , data = string . split ( b'?' , 1 ) except ( ValueError , IndexError ) : items = ( ) else : items = { } mac = hmac ( secret_key , None , cls . hash_method ) for item in data . split ( b'&' ) : mac . update ( b'|' + item ) if not b'=' in item : items = None break key , value = item . split ( b'=' , 1 ) key = url_unquote_plus ( key . decode ( 'ascii' ) ) try : key = to_native ( key ) except UnicodeError : pass items [ key ] = value try : client_hash = base64 . b64decode ( base64_hash ) except TypeError : items = client_hash = None if items is not None and safe_str_cmp ( client_hash , mac . digest ( ) ) : try : for key , value in iteritems ( items ) : items [ key ] = cls . unquote ( value ) except UnquoteError : items = ( ) else : if '_expires' in items : if time ( ) > items [ '_expires' ] : items = ( ) else : del items [ '_expires' ] else : items = ( ) return cls ( items , secret_key , False ) | Load the secure cookie from a serialized string . |
51,055 | def find_model ( sender , model_name ) : MC = get_mc ( ) model = MC . get ( ( MC . c . model_name == model_name ) & ( MC . c . uuid != '' ) ) if model : model_inst = model . get_instance ( ) orm . set_model ( model_name , model_inst . table_name , appname = __name__ , model_path = '' ) return orm . __models__ . get ( model_name ) | Register new model to ORM |
51,056 | def get_model_fields ( model , add_reserver_flag = True ) : import uliweb . orm as orm fields = [ ] m = { 'type' : 'type_name' , 'hint' : 'hint' , 'default' : 'default' , 'required' : 'required' } m1 = { 'index' : 'index' , 'unique' : 'unique' } for name , prop in model . properties . items ( ) : if name == 'id' : continue d = { } for k , v in m . items ( ) : d [ k ] = getattr ( prop , v ) for k , v in m1 . items ( ) : d [ k ] = bool ( prop . kwargs . get ( v ) ) d [ 'name' ] = prop . fieldname or name d [ 'verbose_name' ] = unicode ( prop . verbose_name ) d [ 'nullable' ] = bool ( prop . kwargs . get ( 'nullable' , orm . __nullable__ ) ) if d [ 'type' ] in ( 'VARCHAR' , 'CHAR' , 'BINARY' , 'VARBINARY' ) : d [ 'max_length' ] = prop . max_length if d [ 'type' ] in ( 'Reference' , 'OneToOne' , 'ManyToMany' ) : d [ 'reference_class' ] = prop . reference_class d [ 'collection_name' ] = prop . _collection_name d [ 'server_default' ] = prop . kwargs . get ( 'server_default' ) d [ '_reserved' ] = True fields . append ( d ) return fields | Creating fields suit for model_config id will be skipped . |
51,057 | def get_model_indexes ( model , add_reserver_flag = True ) : import uliweb . orm as orm from sqlalchemy . engine . reflection import Inspector indexes = [ ] engine = model . get_engine ( ) . engine insp = Inspector . from_engine ( engine ) for index in insp . get_indexes ( model . tablename ) : d = { } d [ 'name' ] = index [ 'name' ] d [ 'unique' ] = index [ 'unique' ] d [ 'fields' ] = index [ 'column_names' ] if add_reserver_flag : d [ '_reserved' ] = True indexes . append ( d ) return indexes | Creating indexes suit for model_config . |
51,058 | def timeit ( output ) : b = time . time ( ) yield print output , 'time used: %.3fs' % ( time . time ( ) - b ) | If output is string then print the string and also time used |
51,059 | def host_is_trusted ( hostname , trusted_list ) : if not hostname : return False if isinstance ( trusted_list , string_types ) : trusted_list = [ trusted_list ] def _normalize ( hostname ) : if ':' in hostname : hostname = hostname . rsplit ( ':' , 1 ) [ 0 ] return _encode_idna ( hostname ) hostname = _normalize ( hostname ) for ref in trusted_list : if ref . startswith ( '.' ) : ref = ref [ 1 : ] suffix_match = True else : suffix_match = False ref = _normalize ( ref ) if ref == hostname : return True if suffix_match and hostname . endswith ( '.' + ref ) : return True return False | Checks if a host is trusted against a list . This also takes care of port normalization . |
51,060 | def get_content_length ( environ ) : content_length = environ . get ( 'CONTENT_LENGTH' ) if content_length is not None : try : return max ( 0 , int ( content_length ) ) except ( ValueError , TypeError ) : pass | Returns the content length from the WSGI environment as integer . If it s not available None is returned . |
51,061 | def get_input_stream ( environ , safe_fallback = True ) : stream = environ [ 'wsgi.input' ] content_length = get_content_length ( environ ) if environ . get ( 'wsgi.input_terminated' ) : return stream if content_length is None : return safe_fallback and _empty_stream or stream return LimitedStream ( stream , content_length ) | Returns the input stream from the WSGI environment and wraps it in the most sensible way possible . The stream returned is not the raw WSGI stream in most cases but one that is safe to read from without taking into account the content length . |
51,062 | def get_path_info ( environ , charset = 'utf-8' , errors = 'replace' ) : path = wsgi_get_bytes ( environ . get ( 'PATH_INFO' , '' ) ) return to_unicode ( path , charset , errors , allow_none_charset = True ) | Returns the PATH_INFO from the WSGI environment and properly decodes it . This also takes care about the WSGI decoding dance on Python 3 environments . if the charset is set to None a bytestring is returned . |
51,063 | def pop_path_info ( environ , charset = 'utf-8' , errors = 'replace' ) : path = environ . get ( 'PATH_INFO' ) if not path : return None script_name = environ . get ( 'SCRIPT_NAME' , '' ) old_path = path path = path . lstrip ( '/' ) if path != old_path : script_name += '/' * ( len ( old_path ) - len ( path ) ) if '/' not in path : environ [ 'PATH_INFO' ] = '' environ [ 'SCRIPT_NAME' ] = script_name + path rv = wsgi_get_bytes ( path ) else : segment , path = path . split ( '/' , 1 ) environ [ 'PATH_INFO' ] = '/' + path environ [ 'SCRIPT_NAME' ] = script_name + segment rv = wsgi_get_bytes ( segment ) return to_unicode ( rv , charset , errors , allow_none_charset = True ) | Removes and returns the next segment of PATH_INFO pushing it onto SCRIPT_NAME . Returns None if there is nothing left on PATH_INFO . |
51,064 | def _make_chunk_iter ( stream , limit , buffer_size ) : if isinstance ( stream , ( bytes , bytearray , text_type ) ) : raise TypeError ( 'Passed a string or byte object instead of ' 'true iterator or stream.' ) if not hasattr ( stream , 'read' ) : for item in stream : if item : yield item return if not isinstance ( stream , LimitedStream ) and limit is not None : stream = LimitedStream ( stream , limit ) _read = stream . read while 1 : item = _read ( buffer_size ) if not item : break yield item | Helper for the line and chunk iter functions . |
51,065 | def soap ( func = None , name = None , returns = None , args = None , doc = None , target = 'SOAP' ) : global __soap_functions__ returns = _fix_soap_kwargs ( returns ) args = _fix_soap_kwargs ( args ) if isinstance ( func , str ) and not name : return partial ( soap , name = func , returns = returns , args = args , doc = doc , target = target ) if not func : return partial ( soap , name = name , returns = returns , args = args , doc = doc , target = target ) target_functions = __soap_functions__ . setdefault ( target , { } ) if inspect . isfunction ( func ) : f_name = func . __name__ if name : f_name = name target_functions [ f_name ] = endpoint = ( '.' . join ( [ func . __module__ , func . __name__ ] ) , returns , args , doc ) func . soap_endpoint = ( f_name , endpoint ) elif inspect . isclass ( func ) : if not name : name = func . __name__ for _name in dir ( func ) : f = getattr ( func , _name ) if ( inspect . ismethod ( f ) or inspect . isfunction ( f ) ) and not _name . startswith ( '_' ) : f_name = name + '.' + f . __name__ endpoint = ( '.' . join ( [ func . __module__ , func . __name__ , _name ] ) , returns , args , doc ) if hasattr ( f , 'soap_endpoint' ) : _n , _e = f . soap_endpoint target_functions [ name + '.' + _n ] = endpoint del target_functions [ _n ] else : target_functions [ f_name ] = endpoint else : raise Exception ( "Can't support this type [%r]" % func ) return func | soap supports multiple SOAP function collections it ll save functions to target dict and you can give other target but it should be keep up with SoapView . target definition . |
51,066 | def getJsonData ( self , params ) : try : return eval ( "self." + str ( params [ 'output_id' ] ) + "(params)" ) except AttributeError : df = self . getData ( params ) if df is None : return None return df . to_dict ( orient = 'records' ) | turns the DataFrame returned by getData into a dictionary |
51,067 | def launch_in_notebook ( self , port = 9095 , width = 900 , height = 600 ) : from IPython . lib import backgroundjobs as bg from IPython . display import HTML jobs = bg . BackgroundJobManager ( ) jobs . new ( self . launch , kw = dict ( port = port ) ) frame = HTML ( '<iframe src=http://localhost:{} width={} height={}></iframe>' . format ( port , width , height ) ) return frame | launch the app within an iframe in ipython notebook |
51,068 | def launch ( self , host = "local" , port = 8080 ) : self . root . templateVars [ 'app_bar' ] = self . site_app_bar for fullRoute , _ in self . site_app_bar [ 1 : ] : parent , route = self . get_route ( fullRoute ) parent . __dict__ [ route ] . templateVars [ 'app_bar' ] = self . site_app_bar if host != "local" : cherrypy . server . socket_host = '0.0.0.0' cherrypy . server . socket_port = port cherrypy . quickstart ( self . root ) | Calling the Launch method on a Site object will serve the top node of the cherrypy Root object tree |
51,069 | def exclude_args ( parser , args , excluded_args , target ) : msg = '"%s" option invalid for %s' for argname in excluded_args : if argname not in args : continue if args [ argname ] : optname = "--%s" % argname . replace ( "_" , "-" ) parser . error ( msg % ( optname , target ) ) del args [ argname ] | Delete options that are not appropriate for a following code path ; exit with an error if excluded options were passed in by the user . |
51,070 | def _varLib_finder ( source , directory = "" , ext = "ttf" ) : fname = os . path . splitext ( os . path . basename ( source ) ) [ 0 ] + "." + ext return os . path . join ( directory , fname ) | Finder function to be used with varLib . build to find master TTFs given the filename of the source UFO master as specified in the designspace . It replaces the UFO directory with the one specified in directory argument and replaces the file extension with ext . |
51,071 | def build_master_ufos ( self , glyphs_path , designspace_path = None , master_dir = None , instance_dir = None , family_name = None , mti_source = None , ) : import glyphsLib if master_dir is None : master_dir = self . _output_dir ( "ufo" ) if not os . path . isdir ( master_dir ) : os . mkdir ( master_dir ) if instance_dir is None : instance_dir = self . _output_dir ( "ufo" , is_instance = True ) if not os . path . isdir ( instance_dir ) : os . mkdir ( instance_dir ) font = glyphsLib . GSFont ( glyphs_path ) if designspace_path is not None : designspace_dir = os . path . dirname ( designspace_path ) else : designspace_dir = master_dir instance_dir = os . path . relpath ( instance_dir , designspace_dir ) designspace = glyphsLib . to_designspace ( font , family_name = family_name , instance_dir = instance_dir ) masters = { } for source in designspace . sources : if source . filename in masters : assert source . font is masters [ source . filename ] continue ufo_path = os . path . join ( master_dir , source . filename ) source . path = ufo_path source . font . save ( ufo_path ) masters [ source . filename ] = source . font if designspace_path is None : designspace_path = os . path . join ( master_dir , designspace . filename ) designspace . write ( designspace_path ) if mti_source : self . add_mti_features_to_master_ufos ( mti_source , masters . values ( ) ) return designspace_path | Build UFOs and MutatorMath designspace from Glyphs source . |
51,072 | def remove_overlaps ( self , ufos , glyph_filter = lambda g : len ( g ) ) : from booleanOperations import union , BooleanOperationsError for ufo in ufos : font_name = self . _font_name ( ufo ) logger . info ( "Removing overlaps for " + font_name ) for glyph in ufo : if not glyph_filter ( glyph ) : continue contours = list ( glyph ) glyph . clearContours ( ) try : union ( contours , glyph . getPointPen ( ) ) except BooleanOperationsError : logger . error ( "Failed to remove overlaps for %s: %r" , font_name , glyph . name ) raise | Remove overlaps in UFOs glyphs contours . |
51,073 | def decompose_glyphs ( self , ufos , glyph_filter = lambda g : True ) : for ufo in ufos : logger . info ( "Decomposing glyphs for " + self . _font_name ( ufo ) ) for glyph in ufo : if not glyph . components or not glyph_filter ( glyph ) : continue self . _deep_copy_contours ( ufo , glyph , glyph , Transform ( ) ) glyph . clearComponents ( ) | Move components of UFOs glyphs to their outlines . |
51,074 | def build_ttfs ( self , ufos , ** kwargs ) : self . save_otfs ( ufos , ttf = True , ** kwargs ) | Build OpenType binaries with TrueType outlines . |
51,075 | def build_variable_font ( self , designspace , output_path = None , output_dir = None , master_bin_dir = None , ttf = True , ) : assert not ( output_path and output_dir ) , "mutually exclusive args" ext = "ttf" if ttf else "otf" if hasattr ( designspace , "__fspath__" ) : designspace = designspace . __fspath__ ( ) if isinstance ( designspace , basestring ) : designspace = designspaceLib . DesignSpaceDocument . fromfile ( designspace ) if master_bin_dir is None : master_bin_dir = self . _output_dir ( ext , interpolatable = True ) finder = partial ( _varLib_finder , directory = master_bin_dir ) else : assert all ( isinstance ( s . font , TTFont ) for s in designspace . sources ) finder = lambda s : s if output_path is None : output_path = ( os . path . splitext ( os . path . basename ( designspace . path ) ) [ 0 ] + "-VF" ) output_path = self . _output_path ( output_path , ext , is_variable = True , output_dir = output_dir ) logger . info ( "Building variable font " + output_path ) font , _ , _ = varLib . build ( designspace , finder ) font . save ( output_path ) | Build OpenType variable font from masters in a designspace . |
51,076 | def subset_otf_from_ufo ( self , otf_path , ufo ) : from fontTools import subset ufo_order = makeOfficialGlyphOrder ( ufo ) if ".notdef" not in ufo_order : ufo_order . insert ( 0 , ".notdef" ) ot_order = TTFont ( otf_path ) . getGlyphOrder ( ) assert ot_order [ 0 ] == ".notdef" assert len ( ufo_order ) == len ( ot_order ) for key in ( KEEP_GLYPHS_NEW_KEY , KEEP_GLYPHS_OLD_KEY ) : keep_glyphs_list = ufo . lib . get ( key ) if keep_glyphs_list is not None : keep_glyphs = set ( keep_glyphs_list ) break else : keep_glyphs = None include = [ ] for source_name , binary_name in zip ( ufo_order , ot_order ) : if keep_glyphs and source_name not in keep_glyphs : continue if source_name in ufo : exported = ufo [ source_name ] . lib . get ( GLYPH_EXPORT_KEY , True ) if not exported : continue include . append ( binary_name ) opt = subset . Options ( ) opt . name_IDs = [ "*" ] opt . name_legacy = True opt . name_languages = [ "*" ] opt . layout_features = [ "*" ] opt . notdef_outline = True opt . recalc_bounds = True opt . recalc_timestamp = True opt . canonical_order = True opt . glyph_names = True font = subset . load_font ( otf_path , opt , lazy = False ) subsetter = subset . Subsetter ( options = opt ) subsetter . populate ( glyphs = include ) subsetter . subset ( font ) subset . save_font ( font , otf_path , opt ) | Subset a font using export flags set by glyphsLib . |
51,077 | def run_from_glyphs ( self , glyphs_path , designspace_path = None , master_dir = None , instance_dir = None , family_name = None , mti_source = None , ** kwargs ) : logger . info ( "Building master UFOs and designspace from Glyphs source" ) designspace_path = self . build_master_ufos ( glyphs_path , designspace_path = designspace_path , master_dir = master_dir , instance_dir = instance_dir , family_name = family_name , mti_source = mti_source , ) self . run_from_designspace ( designspace_path , ** kwargs ) | Run toolchain from Glyphs source . |
51,078 | def interpolate_instance_ufos ( self , designspace , include = None , round_instances = False , expand_features_to_instances = False , ) : from glyphsLib . interpolation import apply_instance_data from mutatorMath . ufo . document import DesignSpaceDocumentReader if any ( source . layerName is not None for source in designspace . sources ) : raise FontmakeError ( "MutatorMath doesn't support DesignSpace sources with 'layer' " "attribute" ) builder = DesignSpaceDocumentReader ( designspace . path , ufoVersion = 3 , roundGeometry = round_instances , verbose = True ) logger . info ( "Interpolating master UFOs from designspace" ) if include is not None : instances = self . _search_instances ( designspace , pattern = include ) for instance_name in instances : builder . readInstance ( ( "name" , instance_name ) ) filenames = set ( instances . values ( ) ) else : builder . readInstances ( ) filenames = None logger . info ( "Applying instance data from designspace" ) instance_ufos = apply_instance_data ( designspace , include_filenames = filenames ) if expand_features_to_instances : logger . debug ( "Expanding features to instance UFOs" ) master_source = next ( ( s for s in designspace . sources if s . copyFeatures ) , None ) if not master_source : raise ValueError ( "No source is designated as the master for features." ) else : master_source_font = builder . sources [ master_source . name ] [ 0 ] master_source_features = parseLayoutFeatures ( master_source_font ) . asFea ( ) for instance_ufo in instance_ufos : instance_ufo . features . text = master_source_features instance_ufo . save ( ) return instance_ufos | Interpolate master UFOs with MutatorMath and return instance UFOs . |
51,079 | def run_from_ufos ( self , ufos , output = ( ) , ** kwargs ) : if set ( output ) == { "ufo" } : return ufo_paths = [ ] if isinstance ( ufos , basestring ) : ufo_paths = glob . glob ( ufos ) ufos = [ Font ( x ) for x in ufo_paths ] elif isinstance ( ufos , list ) : ufos = [ Font ( x ) if isinstance ( x , basestring ) else x for x in ufos ] ufo_paths = [ x . path for x in ufos ] else : raise FontmakeError ( "UFOs parameter is neither a defcon.Font object, a path or a glob, " "nor a list of any of these." , ufos , ) need_reload = False if "otf" in output : self . build_otfs ( ufos , ** kwargs ) need_reload = True if "ttf" in output : if need_reload : ufos = [ Font ( path ) for path in ufo_paths ] self . build_ttfs ( ufos , ** kwargs ) need_reload = True | Run toolchain from UFO sources . |
51,080 | def _font_name ( self , ufo ) : family_name = ( ufo . info . familyName . replace ( " " , "" ) if ufo . info . familyName is not None else "None" ) style_name = ( ufo . info . styleName . replace ( " " , "" ) if ufo . info . styleName is not None else "None" ) return "{}-{}" . format ( family_name , style_name ) | Generate a postscript - style font name . |
51,081 | def _output_dir ( self , ext , is_instance = False , interpolatable = False , autohinted = False , is_variable = False , ) : assert not ( is_variable and any ( [ is_instance , interpolatable ] ) ) if is_variable : dir_prefix = "variable_" elif is_instance : dir_prefix = "instance_" else : dir_prefix = "master_" dir_suffix = "_interpolatable" if interpolatable else "" output_dir = dir_prefix + ext + dir_suffix if autohinted : output_dir = os . path . join ( "autohinted" , output_dir ) return output_dir | Generate an output directory . |
51,082 | def _output_path ( self , ufo_or_font_name , ext , is_instance = False , interpolatable = False , autohinted = False , is_variable = False , output_dir = None , suffix = None , ) : if isinstance ( ufo_or_font_name , basestring ) : font_name = ufo_or_font_name elif ufo_or_font_name . path : font_name = os . path . splitext ( os . path . basename ( os . path . normpath ( ufo_or_font_name . path ) ) ) [ 0 ] else : font_name = self . _font_name ( ufo_or_font_name ) if output_dir is None : output_dir = self . _output_dir ( ext , is_instance , interpolatable , autohinted , is_variable ) if not os . path . exists ( output_dir ) : os . makedirs ( output_dir ) if suffix : return os . path . join ( output_dir , "{}-{}.{}" . format ( font_name , suffix , ext ) ) else : return os . path . join ( output_dir , "{}.{}" . format ( font_name , ext ) ) | Generate output path for a font file with given extension . |
51,083 | def _designspace_locations ( self , designspace ) : maps = [ ] for elements in ( designspace . sources , designspace . instances ) : location_map = { } for element in elements : path = _normpath ( element . path ) location_map [ path ] = element . location maps . append ( location_map ) return maps | Map font filenames to their locations in a designspace . |
51,084 | def _closest_location ( self , location_map , target ) : def dist ( a , b ) : return math . sqrt ( sum ( ( a [ k ] - b [ k ] ) ** 2 for k in a . keys ( ) ) ) paths = iter ( location_map . keys ( ) ) closest = next ( paths ) closest_dist = dist ( target , location_map [ closest ] ) for path in paths : cur_dist = dist ( target , location_map [ path ] ) if cur_dist < closest_dist : closest = path closest_dist = cur_dist return closest | Return path of font whose location is closest to target . |
51,085 | def ttfautohint ( in_file , out_file , args = None , ** kwargs ) : arg_list = [ "ttfautohint" ] file_args = [ in_file , out_file ] if args is not None : if kwargs : raise TypeError ( "Should not provide both cmd args and kwargs." ) rv = subprocess . call ( arg_list + args . split ( ) + file_args ) if rv != 0 : raise TTFAError ( rv ) return boolean_options = ( "debug" , "composites" , "dehint" , "help" , "ignore_restrictions" , "detailed_info" , "no_info" , "adjust_subglyphs" , "symbol" , "ttfa_table" , "verbose" , "version" , "windows_compatibility" , ) other_options = ( "default_script" , "fallback_script" , "family_suffix" , "hinting_limit" , "fallback_stem_width" , "hinting_range_min" , "control_file" , "hinting_range_max" , "strong_stem_width" , "increase_x_height" , "x_height_snapping_exceptions" , ) for option in boolean_options : if kwargs . pop ( option , False ) : arg_list . append ( "--" + option . replace ( "_" , "-" ) ) for option in other_options : arg = kwargs . pop ( option , None ) if arg is not None : arg_list . append ( "--{}={}" . format ( option . replace ( "_" , "-" ) , arg ) ) if kwargs : raise TypeError ( "Unexpected argument(s): " + ", " . join ( kwargs . keys ( ) ) ) rv = subprocess . call ( arg_list + file_args ) if rv != 0 : raise TTFAError ( rv ) | Thin wrapper around the ttfautohint command line tool . |
51,086 | def _default_json_default ( obj ) : if isinstance ( obj , ( datetime . datetime , datetime . date , datetime . time ) ) : return obj . isoformat ( ) else : return str ( obj ) | Coerce everything to strings . All objects representing time get output as ISO8601 . |
51,087 | def format ( self , record ) : fields = record . __dict__ . copy ( ) if isinstance ( record . msg , dict ) : fields . update ( record . msg ) fields . pop ( 'msg' ) msg = "" else : msg = record . getMessage ( ) try : msg = msg . format ( ** fields ) except ( KeyError , IndexError , ValueError ) : pass except : msg = msg if 'msg' in fields : fields . pop ( 'msg' ) if 'exc_info' in fields : if fields [ 'exc_info' ] : formatted = tb . format_exception ( * fields [ 'exc_info' ] ) fields [ 'exception' ] = formatted fields . pop ( 'exc_info' ) if 'exc_text' in fields and not fields [ 'exc_text' ] : fields . pop ( 'exc_text' ) logr = self . defaults . copy ( ) logr . update ( { '@message' : msg , '@timestamp' : datetime . datetime . utcnow ( ) . strftime ( '%Y-%m-%dT%H:%M:%S.%fZ' ) , '@source_host' : self . source_host , '@fields' : self . _build_fields ( logr , fields ) } ) return json . dumps ( logr , default = self . json_default , cls = self . json_cls ) | Format a log record to JSON if the message is a dict assume an empty message and use the dict as additional fields . |
51,088 | def _build_fields ( self , defaults , fields ) : return dict ( list ( defaults . get ( '@fields' , { } ) . items ( ) ) + list ( fields . items ( ) ) ) | Return provided fields including any in defaults |
51,089 | def _authenticate ( self ) : self . cleanup_headers ( ) url = LOGIN_ENDPOINT data = self . query ( url , method = 'POST' , extra_params = { 'email' : self . __username , 'password' : self . __password } ) if isinstance ( data , dict ) and data . get ( 'success' ) : data = data . get ( 'data' ) self . authenticated = data . get ( 'authenticated' ) self . country_code = data . get ( 'countryCode' ) self . date_created = data . get ( 'dateCreated' ) self . __token = data . get ( 'token' ) self . userid = data . get ( 'userId' ) self . __headers [ 'Authorization' ] = self . __token | Authenticate user and generate token . |
51,090 | def cleanup_headers ( self ) : headers = { 'Content-Type' : 'application/json' } headers [ 'Authorization' ] = self . __token self . __headers = headers self . __params = { } | Reset the headers and params . |
51,091 | def query ( self , url , method = 'GET' , extra_params = None , extra_headers = None , retry = 3 , raw = False , stream = False ) : response = None loop = 0 self . cleanup_headers ( ) while loop <= retry : if extra_params : params = self . __params params . update ( extra_params ) else : params = self . __params _LOGGER . debug ( "Params: %s" , params ) if extra_headers : headers = self . __headers headers . update ( extra_headers ) else : headers = self . __headers _LOGGER . debug ( "Headers: %s" , headers ) _LOGGER . debug ( "Querying %s on attempt: %s/%s" , url , loop , retry ) loop += 1 req = None if method == 'GET' : req = self . session . get ( url , headers = headers , stream = stream ) elif method == 'PUT' : req = self . session . put ( url , json = params , headers = headers ) elif method == 'POST' : req = self . session . post ( url , json = params , headers = headers ) if req and ( req . status_code == 200 ) : if raw : _LOGGER . debug ( "Required raw object." ) response = req else : response = req . json ( ) break return response | Return a JSON object or raw session . |
51,092 | def devices ( self ) : if self . _all_devices : return self . _all_devices self . _all_devices = { } self . _all_devices [ 'cameras' ] = [ ] self . _all_devices [ 'base_station' ] = [ ] url = DEVICES_ENDPOINT data = self . query ( url ) for device in data . get ( 'data' ) : name = device . get ( 'deviceName' ) if ( ( device . get ( 'deviceType' ) == 'camera' or device . get ( 'deviceType' ) == 'arloq' or device . get ( 'deviceType' ) == 'arloqs' ) and device . get ( 'state' ) == 'provisioned' ) : camera = ArloCamera ( name , device , self ) self . _all_devices [ 'cameras' ] . append ( camera ) if ( device . get ( 'state' ) == 'provisioned' and ( device . get ( 'deviceType' ) == 'basestation' or device . get ( 'modelId' ) == 'ABC1000' ) ) : base = ArloBaseStation ( name , device , self . __token , self ) self . _all_devices [ 'base_station' ] . append ( base ) return self . _all_devices | Return all devices on Arlo account . |
51,093 | def lookup_camera_by_id ( self , device_id ) : camera = list ( filter ( lambda cam : cam . device_id == device_id , self . cameras ) ) [ 0 ] if camera : return camera return None | Return camera object by device_id . |
51,094 | def refresh_attributes ( self , name ) : url = DEVICES_ENDPOINT response = self . query ( url ) if not response or not isinstance ( response , dict ) : return None for device in response . get ( 'data' ) : if device . get ( 'deviceName' ) == name : return device return None | Refresh attributes from a given Arlo object . |
51,095 | def update ( self , update_cameras = False , update_base_station = False ) : self . _authenticate ( ) if update_cameras : url = DEVICES_ENDPOINT response = self . query ( url ) if not response or not isinstance ( response , dict ) : return for camera in self . cameras : for dev_info in response . get ( 'data' ) : if dev_info . get ( 'deviceName' ) == camera . name : _LOGGER . debug ( "Refreshing %s attributes" , camera . name ) camera . attrs = dev_info camera . make_video_cache ( ) if update_base_station : for base in self . base_stations : base . update ( ) | Refresh object . |
51,096 | def load ( self , days = PRELOAD_DAYS , only_cameras = None , date_from = None , date_to = None , limit = None ) : videos = [ ] url = LIBRARY_ENDPOINT if not ( date_from and date_to ) : now = datetime . today ( ) date_from = ( now - timedelta ( days = days ) ) . strftime ( '%Y%m%d' ) date_to = now . strftime ( '%Y%m%d' ) params = { 'dateFrom' : date_from , 'dateTo' : date_to } data = self . _session . query ( url , method = 'POST' , extra_params = params ) . get ( 'data' ) all_cameras = self . _session . cameras for video in data : srccam = list ( filter ( lambda cam : cam . device_id == video . get ( 'deviceId' ) , all_cameras ) ) [ 0 ] if only_cameras and not isinstance ( only_cameras , list ) : only_cameras = [ ( only_cameras ) ] if only_cameras : if list ( filter ( lambda cam : cam . device_id == srccam . device_id , list ( only_cameras ) ) ) : videos . append ( ArloVideo ( video , srccam , self . _session ) ) else : videos . append ( ArloVideo ( video , srccam , self . _session ) ) if limit : return videos [ : limit ] return videos | Load Arlo videos from the given criteria |
51,097 | def _name ( self ) : return "{0} {1} {2}" . format ( self . _camera . name , pretty_timestamp ( self . created_at ) , self . _attrs . get ( 'mediaDuration' ) ) | Define object name . |
51,098 | def created_at_pretty ( self , date_format = None ) : if date_format : return pretty_timestamp ( self . created_at , date_format = date_format ) return pretty_timestamp ( self . created_at ) | Return pretty timestamp . |
51,099 | def created_today ( self ) : if self . datetime . date ( ) == datetime . today ( ) . date ( ) : return True return False | Return True if created today . |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.