idx int64 0 63k | question stringlengths 61 4.03k | target stringlengths 6 1.23k |
|---|---|---|
54,800 | async def multi_set ( self , pairs , ttl = SENTINEL , dumps_fn = None , namespace = None , _conn = None ) : start = time . monotonic ( ) dumps = dumps_fn or self . _serializer . dumps tmp_pairs = [ ] for key , value in pairs : tmp_pairs . append ( ( self . build_key ( key , namespace = namespace ) , dumps ( value ) ) )... | Stores multiple values in the given keys . |
54,801 | async def delete ( self , key , namespace = None , _conn = None ) : start = time . monotonic ( ) ns_key = self . build_key ( key , namespace = namespace ) ret = await self . _delete ( ns_key , _conn = _conn ) logger . debug ( "DELETE %s %d (%.4f)s" , ns_key , ret , time . monotonic ( ) - start ) return ret | Deletes the given key . |
54,802 | async def exists ( self , key , namespace = None , _conn = None ) : start = time . monotonic ( ) ns_key = self . build_key ( key , namespace = namespace ) ret = await self . _exists ( ns_key , _conn = _conn ) logger . debug ( "EXISTS %s %d (%.4f)s" , ns_key , ret , time . monotonic ( ) - start ) return ret | Check key exists in the cache . |
54,803 | async def expire ( self , key , ttl , namespace = None , _conn = None ) : start = time . monotonic ( ) ns_key = self . build_key ( key , namespace = namespace ) ret = await self . _expire ( ns_key , ttl , _conn = _conn ) logger . debug ( "EXPIRE %s %d (%.4f)s" , ns_key , ret , time . monotonic ( ) - start ) return ret | Set the ttl to the given key . By setting it to 0 it will disable it |
54,804 | async def clear ( self , namespace = None , _conn = None ) : start = time . monotonic ( ) ret = await self . _clear ( namespace , _conn = _conn ) logger . debug ( "CLEAR %s %d (%.4f)s" , namespace , ret , time . monotonic ( ) - start ) return ret | Clears the cache in the cache namespace . If an alternative namespace is given it will clear those ones instead . |
54,805 | async def raw ( self , command , * args , _conn = None , ** kwargs ) : start = time . monotonic ( ) ret = await self . _raw ( command , * args , encoding = self . serializer . encoding , _conn = _conn , ** kwargs ) logger . debug ( "%s (%.4f)s" , command , time . monotonic ( ) - start ) return ret | Send the raw command to the underlying client . Note that by using this CMD you will lose compatibility with other backends . |
54,806 | async def close ( self , * args , _conn = None , ** kwargs ) : start = time . monotonic ( ) ret = await self . _close ( * args , _conn = _conn , ** kwargs ) logger . debug ( "CLOSE (%.4f)s" , time . monotonic ( ) - start ) return ret | Perform any resource clean up necessary to exit the program safely . After closing cmd execution is still possible but you will have to close again before exiting . |
54,807 | def get ( self , alias : str ) : try : return self . _caches [ alias ] except KeyError : pass config = self . get_alias_config ( alias ) cache = _create_cache ( ** deepcopy ( config ) ) self . _caches [ alias ] = cache return cache | Retrieve cache identified by alias . Will return always the same instance |
54,808 | def create ( self , alias = None , cache = None , ** kwargs ) : if alias : config = self . get_alias_config ( alias ) elif cache : warnings . warn ( "Creating a cache with an explicit config is deprecated, use 'aiocache.Cache'" , DeprecationWarning , ) config = { "cache" : cache } else : raise TypeError ( "create call ... | Create a new cache . Either alias or cache params are required . You can use kwargs to pass extra parameters to configure the cache . |
54,809 | async def async_poller ( client , initial_response , deserialization_callback , polling_method ) : try : client = client if isinstance ( client , ServiceClientAsync ) else client . _client except AttributeError : raise ValueError ( "Poller client parameter must be a low-level msrest Service Client or a SDK client." ) r... | Async Poller for long running operations . |
54,810 | def send ( self , request , ** kwargs ) : session = request . context . session old_max_redirects = None if 'max_redirects' in kwargs : warnings . warn ( "max_redirects in operation kwargs is deprecated, use config.redirect_policy instead" , DeprecationWarning ) old_max_redirects = session . max_redirects session . max... | Patch the current session with Request level operation config . |
54,811 | def _request ( self , method , url , params , headers , content , form_content ) : request = ClientRequest ( method , self . format_url ( url ) ) if params : request . format_parameters ( params ) if headers : request . headers . update ( headers ) if "Accept" not in request . headers : _LOGGER . debug ( "Accept header... | Create ClientRequest object . |
54,812 | def format_url ( self , url , ** kwargs ) : url = url . format ( ** kwargs ) parsed = urlparse ( url ) if not parsed . scheme or not parsed . netloc : url = url . lstrip ( '/' ) base = self . config . base_url . format ( ** kwargs ) . rstrip ( '/' ) url = urljoin ( base + '/' , url ) return url | Format request URL with the client base URL unless the supplied URL is already absolute . |
54,813 | def get ( self , url , params = None , headers = None , content = None , form_content = None ) : request = self . _request ( 'GET' , url , params , headers , content , form_content ) request . method = 'GET' return request | Create a GET request object . |
54,814 | def put ( self , url , params = None , headers = None , content = None , form_content = None ) : request = self . _request ( 'PUT' , url , params , headers , content , form_content ) return request | Create a PUT request object . |
54,815 | def send_formdata ( self , request , headers = None , content = None , ** config ) : request . headers = headers request . add_formdata ( content ) return self . send ( request , ** config ) | Send data as a multipart form - data request . We only deal with file - like objects or strings at this point . The requests is not yet streamed . |
54,816 | def add_header ( self , header , value ) : warnings . warn ( "Private attribute _client.add_header is deprecated. Use config.headers instead." , DeprecationWarning ) self . config . headers [ header ] = value | Add a persistent header - this header will be applied to all requests sent during the current client session . |
54,817 | def signed_session ( self , session = None ) : session = super ( ApiKeyCredentials , self ) . signed_session ( session ) session . headers . update ( self . in_headers ) try : session . params . update ( self . in_query ) except AttributeError : raise ValueError ( "session.params must be a dict to be used in ApiKeyCred... | Create requests session with ApiKey . |
54,818 | def deserialize_from_text ( cls , data , content_type = None ) : if hasattr ( data , 'read' ) : data = cast ( IO , data ) . read ( ) if isinstance ( data , bytes ) : data_as_str = data . decode ( encoding = 'utf-8-sig' ) else : data_as_str = cast ( str , data ) data_as_str = data_as_str . lstrip ( _BOM ) if content_typ... | Decode data according to content - type . |
54,819 | def deserialize_from_http_generics ( cls , body_bytes , headers ) : content_type = None if 'content-type' in headers : content_type = headers [ 'content-type' ] . split ( ";" ) [ 0 ] . strip ( ) . lower ( ) else : content_type = "application/json" if body_bytes : return cls . deserialize_from_text ( body_bytes , conten... | Deserialize from HTTP response . |
54,820 | def on_response ( self , request , response , ** kwargs ) : if kwargs . get ( "stream" , True ) : return http_response = response . http_response response . context [ self . CONTEXT_NAME ] = self . deserialize_from_http_generics ( http_response . text ( ) , http_response . headers ) | Extract data from the body of a REST response object . |
54,821 | def add_headers ( self , header_dict ) : if not self . response : return for name , data_type in header_dict . items ( ) : value = self . response . headers . get ( name ) value = self . _deserialize ( data_type , value ) self . headers [ name ] = value | Deserialize a specific header . |
54,822 | def log_request ( _ , request , * _args , ** _kwargs ) : if not _LOGGER . isEnabledFor ( logging . DEBUG ) : return try : _LOGGER . debug ( "Request URL: %r" , request . url ) _LOGGER . debug ( "Request method: %r" , request . method ) _LOGGER . debug ( "Request headers:" ) for header , value in request . headers . ite... | Log a client request . |
54,823 | def log_response ( _ , _request , response , * _args , ** kwargs ) : if not _LOGGER . isEnabledFor ( logging . DEBUG ) : return None try : _LOGGER . debug ( "Response status: %r" , response . status_code ) _LOGGER . debug ( "Response headers:" ) for res_header , value in response . headers . items ( ) : _LOGGER . debug... | Log a server response . |
54,824 | def _clear_config ( self ) : for section in self . _config . sections ( ) : self . _config . remove_section ( section ) | Clearout config object in memory . |
54,825 | def format_parameters ( self , params ) : query = urlparse ( self . url ) . query if query : self . url = self . url . partition ( '?' ) [ 0 ] existing_params = { p [ 0 ] : p [ - 1 ] for p in [ p . partition ( '=' ) for p in query . split ( '&' ) ] } params . update ( existing_params ) query_params = [ "{}={}" . format... | Format parameters into a valid query string . It s assumed all parameters have already been quoted as valid URL strings . |
54,826 | def _format_data ( data ) : if hasattr ( data , 'read' ) : data = cast ( IO , data ) data_name = None try : if data . name [ 0 ] != '<' and data . name [ - 1 ] != '>' : data_name = os . path . basename ( data . name ) except ( AttributeError , TypeError ) : pass return ( data_name , data , "application/octet-stream" ) ... | Format field data according to whether it is a stream or a string for a form - data request . |
54,827 | def add_formdata ( self , content = None ) : if content is None : content = { } content_type = self . headers . pop ( 'Content-Type' , None ) if self . headers else None if content_type and content_type . lower ( ) == 'application/x-www-form-urlencoded' : self . data = { f : d for f , d in content . items ( ) if d is n... | Add data as a multipart form - data request to the request . |
54,828 | def raise_with_traceback ( exception , message = "" , * args , ** kwargs ) : exc_type , exc_value , exc_traceback = sys . exc_info ( ) exc_msg = "{}, {}: {}" . format ( message , exc_type . __name__ , exc_value ) error = exception ( exc_msg , * args , ** kwargs ) try : raise error . with_traceback ( exc_traceback ) exc... | Raise exception with a specified traceback . |
54,829 | def _patch_redirect ( session ) : def enforce_http_spec ( resp , request ) : if resp . status_code in ( 301 , 302 ) and request . method not in [ 'GET' , 'HEAD' ] : return False return True redirect_logic = session . resolve_redirects def wrapped_redirect ( resp , req , ** kwargs ) : attempt = enforce_http_spec ( resp ... | Whether redirect policy should be applied based on status code . |
54,830 | def _init_session ( self , session ) : _patch_redirect ( session ) max_retries = self . config . retry_policy ( ) for protocol in self . _protocols : session . adapters [ protocol ] . max_retries = max_retries | Init session level configuration of requests . |
54,831 | def _configure_send ( self , request , ** kwargs ) : requests_kwargs = { } session = kwargs . pop ( 'session' , self . session ) if session is not self . session : self . _init_session ( session ) session . max_redirects = int ( self . config . redirect_policy ( ) ) session . trust_env = bool ( self . config . proxies ... | Configure the kwargs to use with requests . |
54,832 | def full_restapi_key_transformer ( key , attr_desc , value ) : keys = _FLATTEN . split ( attr_desc [ 'key' ] ) return ( [ _decode_attribute_map_key ( k ) for k in keys ] , value ) | A key transformer that returns the full RestAPI key path . |
54,833 | def last_restapi_key_transformer ( key , attr_desc , value ) : key , value = full_restapi_key_transformer ( key , attr_desc , value ) return ( key [ - 1 ] , value ) | A key transformer that returns the last RestAPI key . |
54,834 | def _create_xml_node ( tag , prefix = None , ns = None ) : if prefix and ns : ET . register_namespace ( prefix , ns ) if ns : return ET . Element ( "{" + ns + "}" + tag ) else : return ET . Element ( tag ) | Create a XML node . |
54,835 | def _create_xml_node ( cls ) : try : xml_map = cls . _xml_map except AttributeError : raise ValueError ( "This model has no XML definition" ) return _create_xml_node ( xml_map . get ( 'name' , cls . __name__ ) , xml_map . get ( "prefix" , None ) , xml_map . get ( "ns" , None ) ) | Create XML node from _xml_map . |
54,836 | def validate ( self ) : validation_result = [ ] for attr_name , value in [ ( attr , getattr ( self , attr ) ) for attr in self . _attribute_map ] : attr_desc = self . _attribute_map [ attr_name ] if attr_name == "additional_properties" and attr_desc [ "key" ] == '' : continue attr_type = attr_desc [ 'type' ] try : debu... | Validate this model recursively and return a list of ValidationError . |
54,837 | def serialize ( self , keep_readonly = False ) : serializer = Serializer ( self . _infer_class_models ( ) ) return serializer . _serialize ( self , keep_readonly = keep_readonly ) | Return the JSON that would be sent to azure from this model . |
54,838 | def as_dict ( self , keep_readonly = True , key_transformer = attribute_transformer ) : serializer = Serializer ( self . _infer_class_models ( ) ) return serializer . _serialize ( self , key_transformer = key_transformer , keep_readonly = keep_readonly ) | Return a dict that can be JSONify using json . dump . |
54,839 | def deserialize ( cls , data , content_type = None ) : deserializer = Deserializer ( cls . _infer_class_models ( ) ) return deserializer ( cls . __name__ , data , content_type = content_type ) | Parse a str using the RestAPI syntax and return a model . |
54,840 | def from_dict ( cls , data , key_extractors = None , content_type = None ) : deserializer = Deserializer ( cls . _infer_class_models ( ) ) deserializer . key_extractors = [ rest_key_case_insensitive_extractor , attribute_key_case_insensitive_extractor , last_rest_key_case_insensitive_extractor ] if key_extractors is No... | Parse a dict using given key extractor return a model . |
54,841 | def _classify ( cls , response , objects ) : for subtype_key in cls . __dict__ . get ( '_subtype_map' , { } ) . keys ( ) : subtype_value = None rest_api_response_key = cls . _get_rest_key_parts ( subtype_key ) [ - 1 ] subtype_value = response . pop ( rest_api_response_key , None ) or response . pop ( subtype_key , None... | Check the class _subtype_map for any child classes . We want to ignore any inherited _subtype_maps . Remove the polymorphic key from the initial data . |
54,842 | def body ( self , data , data_type , ** kwargs ) : if data is None : raise ValidationError ( "required" , "body" , True ) internal_data_type = data_type . strip ( '[]{}' ) internal_data_type = self . dependencies . get ( internal_data_type , None ) if internal_data_type and not isinstance ( internal_data_type , Enum ) ... | Serialize data intended for a request body . |
54,843 | def url ( self , name , data , data_type , ** kwargs ) : if self . client_side_validation : data = self . validate ( data , name , required = True , ** kwargs ) try : output = self . serialize_data ( data , data_type , ** kwargs ) if data_type == 'bool' : output = json . dumps ( output ) if kwargs . get ( 'skip_quote' ... | Serialize data intended for a URL path . |
54,844 | def header ( self , name , data , data_type , ** kwargs ) : if self . client_side_validation : data = self . validate ( data , name , required = True , ** kwargs ) try : if data_type in [ '[str]' ] : data = [ "" if d is None else d for d in data ] output = self . serialize_data ( data , data_type , ** kwargs ) if data_... | Serialize data intended for a request header . |
54,845 | def validate ( cls , data , name , ** kwargs ) : required = kwargs . get ( 'required' , False ) if required and data is None : raise ValidationError ( "required" , name , True ) elif data is None : return elif kwargs . get ( 'readonly' ) : return try : for key , value in kwargs . items ( ) : validator = cls . validatio... | Validate that a piece of data meets certain conditions |
54,846 | def serialize_data ( self , data , data_type , ** kwargs ) : if data is None : raise ValueError ( "No value for given attribute" ) try : if data_type in self . basic_types . values ( ) : return self . serialize_basic ( data , data_type , ** kwargs ) elif data_type in self . serialize_type : return self . serialize_type... | Serialize generic data according to supplied data type . |
54,847 | def serialize_basic ( self , data , data_type , ** kwargs ) : custom_serializer = self . _get_custom_serializers ( data_type , ** kwargs ) if custom_serializer : return custom_serializer ( data ) if data_type == 'str' : return self . serialize_unicode ( data ) return eval ( data_type ) ( data ) | Serialize basic builting data type . Serializes objects to str int float or bool . |
54,848 | def serialize_unicode ( self , data ) : try : return data . value except AttributeError : pass try : if isinstance ( data , unicode ) : return data . encode ( encoding = 'utf-8' ) except NameError : return str ( data ) else : return str ( data ) | Special handling for serializing unicode strings in Py2 . Encode to UTF - 8 if unicode otherwise handle as a str . |
54,849 | def serialize_iter ( self , data , iter_type , div = None , ** kwargs ) : if isinstance ( data , str ) : raise SerializationError ( "Refuse str type as a valid iter type." ) serialization_ctxt = kwargs . get ( "serialization_ctxt" , { } ) serialized = [ ] for d in data : try : serialized . append ( self . serialize_dat... | Serialize iterable . |
54,850 | def serialize_dict ( self , attr , dict_type , ** kwargs ) : serialization_ctxt = kwargs . get ( "serialization_ctxt" , { } ) serialized = { } for key , value in attr . items ( ) : try : serialized [ self . serialize_unicode ( key ) ] = self . serialize_data ( value , dict_type , ** kwargs ) except ValueError : seriali... | Serialize a dictionary of objects . |
54,851 | def serialize_base64 ( attr , ** kwargs ) : encoded = b64encode ( attr ) . decode ( 'ascii' ) return encoded . strip ( '=' ) . replace ( '+' , '-' ) . replace ( '/' , '_' ) | Serialize str into base - 64 string . |
54,852 | def serialize_date ( attr , ** kwargs ) : if isinstance ( attr , str ) : attr = isodate . parse_date ( attr ) t = "{:04}-{:02}-{:02}" . format ( attr . year , attr . month , attr . day ) return t | Serialize Date object into ISO - 8601 formatted string . |
54,853 | def serialize_duration ( attr , ** kwargs ) : if isinstance ( attr , str ) : attr = isodate . parse_duration ( attr ) return isodate . duration_isoformat ( attr ) | Serialize TimeDelta object into ISO - 8601 formatted string . |
54,854 | def serialize_rfc ( attr , ** kwargs ) : try : if not attr . tzinfo : _LOGGER . warning ( "Datetime with no tzinfo will be considered UTC." ) utc = attr . utctimetuple ( ) except AttributeError : raise TypeError ( "RFC1123 object must be valid Datetime object." ) return "{}, {:02} {} {:04} {:02}:{:02}:{:02} GMT" . form... | Serialize Datetime object into RFC - 1123 formatted string . |
54,855 | def serialize_iso ( attr , ** kwargs ) : if isinstance ( attr , str ) : attr = isodate . parse_datetime ( attr ) try : if not attr . tzinfo : _LOGGER . warning ( "Datetime with no tzinfo will be considered UTC." ) utc = attr . utctimetuple ( ) if utc . tm_year > 9999 or utc . tm_year < 1 : raise OverflowError ( "Hit ma... | Serialize Datetime object into ISO - 8601 formatted string . |
54,856 | def _deserialize ( self , target_obj , data ) : if hasattr ( data , "_attribute_map" ) : constants = [ name for name , config in getattr ( data , '_validation' , { } ) . items ( ) if config . get ( 'constant' ) ] try : for attr , mapconfig in data . _attribute_map . items ( ) : if attr in constants : continue value = g... | Call the deserializer on a model . |
54,857 | def _classify_target ( self , target , data ) : if target is None : return None , None if isinstance ( target , basestring ) : try : target = self . dependencies [ target ] except KeyError : return target , target try : target = target . _classify ( data , self . dependencies ) except AttributeError : pass return targe... | Check to see whether the deserialization target object can be classified into a subclass . Once classification has been determined initialize object . |
54,858 | def _unpack_content ( raw_data , content_type = None ) : from . pipeline . universal import RawDeserializer context = getattr ( raw_data , "context" , { } ) if context : if RawDeserializer . CONTEXT_NAME in context : return context [ RawDeserializer . CONTEXT_NAME ] raise ValueError ( "This pipeline didn't have the Raw... | Extract the correct structure for deserialization . |
54,859 | def _instantiate_model ( self , response , attrs , additional_properties = None ) : if callable ( response ) : subtype = getattr ( response , '_subtype_map' , { } ) try : readonly = [ k for k , v in response . _validation . items ( ) if v . get ( 'readonly' ) ] const = [ k for k , v in response . _validation . items ( ... | Instantiate a response model passing in deserialized args . |
54,860 | def deserialize_data ( self , data , data_type ) : if data is None : return data try : if not data_type : return data if data_type in self . basic_types . values ( ) : return self . deserialize_basic ( data , data_type ) if data_type in self . deserialize_type : if isinstance ( data , self . deserialize_expected_types ... | Process data for deserialization according to data type . |
54,861 | def deserialize_iter ( self , attr , iter_type ) : if attr is None : return None if isinstance ( attr , ET . Element ) : attr = list ( attr ) if not isinstance ( attr , ( list , set ) ) : raise DeserializationError ( "Cannot deserialize as [{}] an object of type {}" . format ( iter_type , type ( attr ) ) ) return [ sel... | Deserialize an iterable . |
54,862 | def deserialize_dict ( self , attr , dict_type ) : if isinstance ( attr , list ) : return { x [ 'key' ] : self . deserialize_data ( x [ 'value' ] , dict_type ) for x in attr } if isinstance ( attr , ET . Element ) : attr = { el . tag : el . text for el in attr } return { k : self . deserialize_data ( v , dict_type ) fo... | Deserialize a dictionary . |
54,863 | def deserialize_object ( self , attr , ** kwargs ) : if attr is None : return None if isinstance ( attr , ET . Element ) : return attr if isinstance ( attr , basestring ) : return self . deserialize_basic ( attr , 'str' ) obj_type = type ( attr ) if obj_type in self . basic_types : return self . deserialize_basic ( att... | Deserialize a generic object . This will be handled as a dictionary . |
54,864 | def deserialize_basic ( self , attr , data_type ) : if isinstance ( attr , ET . Element ) : attr = attr . text if not attr : if data_type == "str" : return '' else : return None if data_type == 'bool' : if attr in [ True , False , 1 , 0 ] : return bool ( attr ) elif isinstance ( attr , basestring ) : if attr . lower ( ... | Deserialize baisc builtin data type from string . Will attempt to convert to str int float and bool . This function will also accept 1 0 true and false as valid bool values . |
54,865 | def deserialize_unicode ( data ) : if isinstance ( data , Enum ) : return data try : if isinstance ( data , unicode ) : return data except NameError : return str ( data ) else : return str ( data ) | Preserve unicode objects in Python 2 otherwise return data as a string . |
54,866 | def deserialize_enum ( data , enum_obj ) : if isinstance ( data , enum_obj ) : return data if isinstance ( data , Enum ) : data = data . value if isinstance ( data , int ) : try : return list ( enum_obj . __members__ . values ( ) ) [ data ] except IndexError : error = "{!r} is not a valid index for enum {!r}" raise Des... | Deserialize string into enum object . |
54,867 | def deserialize_bytearray ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text return bytearray ( b64decode ( attr ) ) | Deserialize string into bytearray . |
54,868 | def deserialize_base64 ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text padding = '=' * ( 3 - ( len ( attr ) + 3 ) % 4 ) attr = attr + padding encoded = attr . replace ( '-' , '+' ) . replace ( '_' , '/' ) return b64decode ( encoded ) | Deserialize base64 encoded string into string . |
54,869 | def deserialize_decimal ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text try : return decimal . Decimal ( attr ) except decimal . DecimalException as err : msg = "Invalid decimal {}" . format ( attr ) raise_with_traceback ( DeserializationError , msg , err ) | Deserialize string into Decimal object . |
54,870 | def deserialize_duration ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text try : duration = isodate . parse_duration ( attr ) except ( ValueError , OverflowError , AttributeError ) as err : msg = "Cannot deserialize duration object." raise_with_traceback ( DeserializationError , msg , err ) else : r... | Deserialize ISO - 8601 formatted string into TimeDelta object . |
54,871 | def deserialize_date ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text if re . search ( r"[^\W\d_]" , attr , re . I + re . U ) : raise DeserializationError ( "Date must have only digits and -. Received: %s" % attr ) return isodate . parse_date ( attr , defaultmonth = None , defaultday = None ) | Deserialize ISO - 8601 formatted string into Date object . |
54,872 | def deserialize_rfc ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text try : date_obj = datetime . datetime . strptime ( attr , "%a, %d %b %Y %H:%M:%S %Z" ) if not date_obj . tzinfo : date_obj = date_obj . replace ( tzinfo = TZ_UTC ) except ValueError as err : msg = "Cannot deserialize to rfc datetim... | Deserialize RFC - 1123 formatted string into Datetime object . |
54,873 | def deserialize_iso ( attr ) : if isinstance ( attr , ET . Element ) : attr = attr . text try : attr = attr . upper ( ) match = Deserializer . valid_date . match ( attr ) if not match : raise ValueError ( "Invalid datetime string: " + attr ) check_decimal = attr . split ( '.' ) if len ( check_decimal ) > 1 : decimal_st... | Deserialize ISO - 8601 formatted string into Datetime object . |
54,874 | def raw ( self ) : raw = ClientRawResponse ( self . current_page , self . _response ) if self . _raw_headers : raw . add_headers ( self . _raw_headers ) return raw | Get current page as ClientRawResponse . |
54,875 | def advance_page ( self ) : if self . next_link is None : raise StopIteration ( "End of paging" ) self . _current_page_iter_index = 0 self . _response = self . _get_next ( self . next_link ) self . _derserializer ( self , self . _response ) return self . current_page | Force moving the cursor to the next azure call . |
54,876 | def _ensureAtomicity ( fn ) : @ ensureScoopStartedProperly def wrapper ( * args , ** kwargs ) : from . import _control _control . execQueue . socket . pumpInfoSocket ( ) for key , value in kwargs . items ( ) : if key in itertools . chain ( * ( elem . keys ( ) for elem in elements . values ( ) ) ) : raise TypeError ( "T... | Ensure atomicity of passed elements on the whole worker pool |
54,877 | def getConst ( name , timeout = 0.1 ) : from . import _control import time timeStamp = time . time ( ) while True : _control . execQueue . socket . pumpInfoSocket ( ) constants = dict ( reduce ( lambda x , y : x + list ( y . items ( ) ) , elements . values ( ) , [ ] ) ) timeoutHappened = time . time ( ) - timeStamp > t... | Get a shared constant . |
54,878 | def launchBootstraps ( ) : global processes worker_amount , verbosity , args = getArgs ( ) was_origin = False if verbosity >= 1 : sys . stderr . write ( "Launching {0} worker(s) using {1}.\n" . format ( worker_amount , os . environ [ 'SHELL' ] if 'SHELL' in os . environ else "an unknown shell" , ) ) sys . stderr . flus... | Launch the bootstrap instances in separate subprocesses |
54,879 | def resolve ( self , s ) : name = s . split ( '.' ) used = name . pop ( 0 ) try : found = self . importer ( used ) for frag in name : used += '.' + frag try : found = getattr ( found , frag ) except AttributeError : self . importer ( used ) found = getattr ( found , frag ) return found except ImportError : e , tb = sys... | Resolve strings to objects using standard import and attribute syntax . |
54,880 | def as_tuple ( self , value ) : if isinstance ( value , list ) : value = tuple ( value ) return value | Utility function which converts lists to tuples . |
54,881 | def configure_formatter ( self , config ) : if '()' in config : factory = config [ '()' ] try : result = self . configure_custom ( config ) except TypeError , te : if "'format'" not in str ( te ) : raise config [ 'fmt' ] = config . pop ( 'format' ) config [ '()' ] = factory result = self . configure_custom ( config ) e... | Configure a formatter from a dictionary . |
54,882 | def configure_filter ( self , config ) : if '()' in config : result = self . configure_custom ( config ) else : name = config . get ( 'name' , '' ) result = logging . Filter ( name ) return result | Configure a filter from a dictionary . |
54,883 | def configure_logger ( self , name , config , incremental = False ) : logger = logging . getLogger ( name ) self . common_logger_config ( logger , config , incremental ) propagate = config . get ( 'propagate' , None ) if propagate is not None : logger . propagate = propagate | Configure a non - root logger from a dictionary . |
54,884 | def configure_root ( self , config , incremental = False ) : root = logging . getLogger ( ) self . common_logger_config ( root , config , incremental ) | Configure a root logger from a dictionary . |
54,885 | def sliceImage ( image , divWidth , divHeight ) : w , h = image . size tiles = [ ] for y in range ( 0 , h - 1 , h / divHeight ) : my = min ( y + h / divHeight , h ) for x in range ( 0 , w - 1 , w / divWidth ) : mx = min ( x + w / divWidth , w ) tiles . append ( image . crop ( ( x , y , mx , my ) ) ) return tiles | Divide the received image in multiple tiles |
54,886 | def resizeTile ( index , size ) : resized = tiles [ index ] . resize ( size , Image . ANTIALIAS ) return sImage ( resized . tostring ( ) , resized . size , resized . mode ) | Apply Antialiasing resizing to tile |
54,887 | def initLogging ( verbosity = 0 , name = "SCOOP" ) : global loggingConfig verbose_levels = { - 2 : "CRITICAL" , - 1 : "ERROR" , 0 : "WARNING" , 1 : "INFO" , 2 : "DEBUG" , 3 : "DEBUG" , 4 : "NOSET" , } log_handlers = { "console" : { "class" : "logging.StreamHandler" , "formatter" : "{name}Formatter" . format ( name = na... | Creates a logger . |
54,888 | def externalHostname ( hosts ) : hostname = hosts [ 0 ] [ 0 ] if hostname in localHostnames and len ( hosts ) > 1 : hostname = socket . getfqdn ( ) . split ( "." ) [ 0 ] try : socket . getaddrinfo ( hostname , None ) except socket . gaierror : raise Exception ( "\nThe first host (containing a broker) is not" " routable... | Ensure external hostname is routable . |
54,889 | def getHosts ( filename = None , hostlist = None ) : if filename : return getHostsFromFile ( filename ) elif hostlist : return getHostsFromList ( hostlist ) elif getEnv ( ) == "SLURM" : return getHostsFromSLURM ( ) elif getEnv ( ) == "PBS" : return getHostsFromPBS ( ) elif getEnv ( ) == "SGE" : return getHostsFromSGE (... | Return a list of hosts depending on the environment |
54,890 | def getHostsFromFile ( filename ) : valid_hostname = r"^[^ /\t=\n]+" workers = r"\d+" hostname_re = re . compile ( valid_hostname ) worker_re = re . compile ( workers ) hosts = [ ] with open ( filename ) as f : for line in f : if re . search ( '[\[\]]' , line ) : hosts = hosts + parseSLURM ( line . strip ( ) ) else : h... | Parse a file to return a list of hosts . |
54,891 | def getHostsFromList ( hostlist ) : if any ( re . search ( '[\[\]]' , x ) for x in hostlist ) : return parseSLURM ( str ( hostlist ) ) hostlist = groupTogether ( hostlist ) retVal = [ ] for key , group in groupby ( hostlist ) : retVal . append ( ( key , len ( list ( group ) ) ) ) return retVal | Return the hosts from the command line |
54,892 | def parseSLURM ( string ) : import subprocess , os hostsstr = subprocess . check_output ( [ "scontrol" , "show" , "hostnames" , string ] ) if sys . version_info . major > 2 : hostsstr = hostsstr . decode ( ) hosts = hostsstr . split ( os . linesep ) hosts = filter ( None , hosts ) hosts = [ ( host , 1 ) for host in hos... | Return a host list from a SLURM string |
54,893 | def getHostsFromPBS ( ) : with open ( os . environ [ "PBS_NODEFILE" ] , 'r' ) as hosts : hostlist = groupTogether ( hosts . read ( ) . split ( ) ) retVal = [ ] for key , group in groupby ( hostlist ) : retVal . append ( ( key , len ( list ( group ) ) ) ) return retVal | Return a host list in a PBS environment |
54,894 | def getHostsFromSGE ( ) : with open ( os . environ [ "PE_HOSTFILE" ] , 'r' ) as hosts : return [ ( host . split ( ) [ 0 ] , int ( host . split ( ) [ 1 ] ) ) for host in hosts ] | Return a host list in a SGE environment |
54,895 | def getWorkerQte ( hosts ) : if "SLURM_NTASKS" in os . environ : return int ( os . environ [ "SLURM_NTASKS" ] ) elif "PBS_NP" in os . environ : return int ( os . environ [ "PBS_NP" ] ) elif "NSLOTS" in os . environ : return int ( os . environ [ "NSLOTS" ] ) else : return sum ( host [ 1 ] for host in hosts ) | Return the number of workers to launch depending on the environment |
54,896 | def functionFactory ( in_code , name , defaults , globals_ , imports ) : def generatedFunction ( ) : pass generatedFunction . __code__ = marshal . loads ( in_code ) generatedFunction . __name__ = name generatedFunction . __defaults = defaults generatedFunction . __globals__ . update ( pickle . loads ( globals_ ) ) for ... | Creates a function at runtime using binary compiled inCode |
54,897 | def makeLambdaPicklable ( lambda_function ) : if isinstance ( lambda_function , type ( lambda : None ) ) and lambda_function . __name__ == '<lambda>' : def __reduce_ex__ ( proto ) : return unpickleLambda , ( marshal . dumps ( lambda_function . __code__ ) , ) lambda_function . __reduce_ex__ = __reduce_ex__ return lambda... | Take input lambda function l and makes it picklable . |
54,898 | def addConnector ( self , wire1 , wire2 ) : if wire1 == wire2 : return if wire1 > wire2 : wire1 , wire2 = wire2 , wire1 try : last_level = self [ - 1 ] except IndexError : self . append ( [ ( wire1 , wire2 ) ] ) return for wires in last_level : if wires [ 1 ] >= wire1 and wires [ 0 ] <= wire2 : self . append ( [ ( wire... | Add a connector between wire1 and wire2 in the network . |
54,899 | def sort ( self , values ) : for level in self : for wire1 , wire2 in level : if values [ wire1 ] > values [ wire2 ] : values [ wire1 ] , values [ wire2 ] = values [ wire2 ] , values [ wire1 ] | Sort the values in - place based on the connectors in the network . |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.