code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) lowercase__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } lowercase__ = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[Any]: for attribute in key.split('.' ): a__: Optional[Any] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: Any = value elif weight_type == "weight_g": a__: Any = value elif weight_type == "weight_v": a__: Optional[int] = value elif weight_type == "bias": a__: Any = value else: a__: List[str] = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[str] = [] a__: Tuple = fairseq_model.state_dict() a__: int = hf_model.feature_extractor a__: str = hf_model.adapter for name, value in fairseq_dict.items(): a__: List[Any] = False if "conv_layers" in name: load_conv_layer( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == 'group' , ) a__: str = True elif any(x in name for x in ['adaptor', 'w2v_encoder.proj.', 'w2v_proj_ln.'] ): load_adapter(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Dict = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: a__: Dict = True if "*" in mapped_key: a__: str = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: List[Any] = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: str = 'weight_g' elif "weight_v" in name: a__: str = 'weight_v' elif "bias" in name: a__: Dict = 'bias' elif "weight" in name: a__: List[str] = 'weight' else: a__: int = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: List[str] = full_name.split('conv_layers.' )[-1] a__: Tuple = name.split('.' ) a__: Optional[int] = int(items[0] ) a__: Optional[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) a__: Tuple = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) a__: Optional[Any] = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) a__: Tuple = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) a__: List[str] = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: Optional[Any] = full_name.split('adaptor.' )[-1] a__: List[str] = name.split('.' ) if items[1].isdigit(): a__: Optional[int] = int(items[1] ) else: a__: Union[str, Any] = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.' a__: Union[str, Any] = value logger.info(F'Adapter proj layer norm bias was initialized from {full_name}.' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.' a__: List[Any] = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.' a__: Optional[Any] = value logger.info(F'Adapter proj layer bias was initialized from {full_name}.' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.' a__: Dict = value logger.info(F'Adapter proj layer weight was initialized from {full_name}.' ) elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F'{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.' a__: List[Any] = value logger.info(F'Adapter layer {layer_id} bias was initialized from {full_name}.' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F'{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.' a__: str = value logger.info(F'Adapter layer {layer_id} bias was initialized from {full_name}.' ) else: unused_weights.append(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__ , a__: str = emb.weight.shape a__: List[str] = nn.Linear(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , bias=_SCREAMING_SNAKE_CASE ) a__: str = emb.weight.data return lin_layer @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) ->int: a__: Optional[int] = WavaVecaConfig.from_pretrained( _SCREAMING_SNAKE_CASE , add_adapter=_SCREAMING_SNAKE_CASE , adapter_stride=_SCREAMING_SNAKE_CASE , adapter_kernel_size=_SCREAMING_SNAKE_CASE , use_auth_token=_SCREAMING_SNAKE_CASE , output_hidden_size=_SCREAMING_SNAKE_CASE , ) a__: List[Any] = MBartConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) # load model a__ , a__ , a__: Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ 'config_yaml': config_yaml_path, 'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path, 'load_pretrained_decoder_from': None, } , ) a__: Optional[Any] = model[0].eval() # load feature extractor a__: Any = WavaVecaFeatureExtractor.from_pretrained(_SCREAMING_SNAKE_CASE , use_auth_token=_SCREAMING_SNAKE_CASE ) # set weights for wav2vec2 encoder a__: Dict = WavaVecaModel(_SCREAMING_SNAKE_CASE ) recursively_load_weights_wavaveca(model.encoder , _SCREAMING_SNAKE_CASE ) # load decoder weights a__: Optional[int] = MBartForCausalLM(_SCREAMING_SNAKE_CASE ) a__ , a__: Dict = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=_SCREAMING_SNAKE_CASE ) logger.warning(F'The following keys are missing when loading the decoder weights: {missing_keys}' ) logger.warning(F'The following keys are unexpected when loading the decoder weights: {unexpected_keys}' ) a__: Dict = SpeechEncoderDecoderModel(encoder=_SCREAMING_SNAKE_CASE , decoder=_SCREAMING_SNAKE_CASE ) a__: List[Any] = False a__: str = MBartaaTokenizer(_SCREAMING_SNAKE_CASE ) tokenizer.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: int = hf_wavavec.config.to_dict() a__: List[Any] = tokenizer.pad_token_id a__: str = tokenizer.bos_token_id a__: str = tokenizer.eos_token_id a__: Any = 'mbart50' a__: Optional[int] = 'wav2vec2' a__: Dict = tokenizer.eos_token_id a__: Tuple = 250004 a__: Any = tokenizer.eos_token_id a__: List[Any] = SpeechEncoderDecoderConfig.from_dict(_SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(_SCREAMING_SNAKE_CASE ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_yaml_path', default=None, type=str, help='Path to yaml file of fine-tuned model') parser.add_argument( '--encoder_config_path', default='facebook/wav2vec2-xls-r-1b', type=str, help='Path to hf encoder wav2vec2 checkpoint config', ) parser.add_argument( '--decoder_config_path', default='facebook/mbart-large-50-one-to-many-mmt', type=str, help='Path to hf decoder checkpoint config', ) parser.add_argument('--add_adapter', default=True, type=bool, help='whethere to add model adapter layers') parser.add_argument('--adapter_stride', default=2, type=int, help='stride of adapter layers') parser.add_argument('--adapter_kernel_size', default=3, type=int, help='kernel size of adapter layers') parser.add_argument('--encoder_output_dim', default=1024, type=int, help='encoder output dim') parser.add_argument('--start_token_id', default=250004, type=int, help='`decoder_start_token_id` of model config') lowercase__ = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
290
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex lowercase__ = logging.getLogger(__name__) class __snake_case : def __init__( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' if not self.initialized: a__: Optional[int] = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Optional[int] = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.retriever.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ , a__: str = self.retriever._main_retrieve(lowercase , lowercase) return doc_ids, retrieved_doc_embeds class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> int: '''simple docstring''' if index is not None and index.is_initialized() and len(lowercase) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ') super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Any = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase) for worker in self.retrieval_workers ]) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' logger.info('initializing retrieval') if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if len(self.retrieval_workers) > 0: # Select a random retrieval actor. a__: int = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] a__ , a__: List[Any] = ray.get(random_worker.retrieve.remote(lowercase , lowercase)) else: a__ , a__: Dict = self._main_retrieve(lowercase , lowercase) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' return super(lowercase , cls).get_tokenizers(lowercase , lowercase , **lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = kwargs.pop('config' , lowercase) or RagConfig.from_pretrained(lowercase , **lowercase) a__: Union[str, Any] = RagTokenizer.from_pretrained(lowercase , config=lowercase) a__: int = rag_tokenizer.question_encoder a__: Any = rag_tokenizer.generator if indexed_dataset is not None: a__: List[Any] = 'custom' a__: Optional[Any] = CustomHFIndex(config.retrieval_vector_size , lowercase) else: a__: Dict = cls._build_index(lowercase) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
290
1
"""simple docstring""" from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 lowercase__ = { # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 2048-bit 14: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 3072-bit 15: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 4096-bit 16: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 6144-bit 17: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 8192-bit 18: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, } class __snake_case : def __init__( self , lowercase = 14) -> None: '''simple docstring''' if group not in primes: raise ValueError('Unsupported Group') a__: List[Any] = primes[group]['prime'] a__: Any = primes[group]['generator'] a__: str = int(hexlify(urandom(32)) , base=16) def lowerCamelCase_ ( self) -> str: '''simple docstring''' return hex(self.__private_key)[2:] def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: int = pow(self.generator , self.__private_key , self.prime) return hex(lowercase)[2:] def lowerCamelCase_ ( self , lowercase) -> bool: '''simple docstring''' return ( 2 <= key <= self.prime - 2 and pow(lowercase , (self.prime - 1) // 2 , self.prime) == 1 ) def lowerCamelCase_ ( self , lowercase) -> str: '''simple docstring''' a__: Dict = int(lowercase , base=16) if not self.is_valid_public_key(lowercase): raise ValueError('Invalid public key') a__: int = pow(lowercase , self.__private_key , self.prime) return shaaaa(str(lowercase).encode()).hexdigest() @staticmethod def lowerCamelCase_ ( lowercase , lowercase) -> bool: '''simple docstring''' return ( 2 <= remote_public_key_str <= prime - 2 and pow(lowercase , (prime - 1) // 2 , lowercase) == 1 ) @staticmethod def lowerCamelCase_ ( lowercase , lowercase , lowercase = 14) -> str: '''simple docstring''' a__: Tuple = int(lowercase , base=16) a__: Tuple = int(lowercase , base=16) a__: Union[str, Any] = primes[group]['prime'] if not DiffieHellman.is_valid_public_key_static(lowercase , lowercase): raise ValueError('Invalid public key') a__: Union[str, Any] = pow(lowercase , lowercase , lowercase) return shaaaa(str(lowercase).encode()).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: int = None if token is not None: a__: Tuple = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Optional[Any] = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: str = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) a__: int = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Dict = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Dict: a__: Dict = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Dict = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: List[Any] = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Optional[int] = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: List[Any] = None if token is not None: a__: Optional[int] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = result.headers['Location'] a__: Optional[int] = requests.get(_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: int = os.path.join(_SCREAMING_SNAKE_CASE , F'{artifact_name}.zip' ) with open(_SCREAMING_SNAKE_CASE , 'wb' ) as fp: fp.write(response.content ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: a__: List[Any] = [] a__: Optional[Any] = [] a__: List[Any] = None with zipfile.ZipFile(_SCREAMING_SNAKE_CASE ) as z: for filename in z.namelist(): if not os.path.isdir(_SCREAMING_SNAKE_CASE ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(_SCREAMING_SNAKE_CASE ) as f: for line in f: a__: Optional[int] = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs a__: Union[str, Any] = line[: line.index(': ' )] a__: Union[str, Any] = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed a__: Optional[int] = line[len('FAILED ' ) :] failed_tests.append(_SCREAMING_SNAKE_CASE ) elif filename == "job_name.txt": a__: Union[str, Any] = line if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( F'`errors` and `failed_tests` should have the same number of elements. Got {len(_SCREAMING_SNAKE_CASE )} for `errors` ' F'and {len(_SCREAMING_SNAKE_CASE )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some' ' problem.' ) a__: Tuple = None if job_name and job_links: a__: Dict = job_links.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # A list with elements of the form (line of error, error, failed test) a__: int = [x + [y] + [job_link] for x, y in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )] return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->str: a__: int = [] a__: Optional[int] = [os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for p in os.listdir(_SCREAMING_SNAKE_CASE ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(_SCREAMING_SNAKE_CASE , job_links=_SCREAMING_SNAKE_CASE ) ) return errors def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Any: a__: str = Counter() counter.update([x[1] for x in logs] ) a__: int = counter.most_common() a__: Any = {} for error, count in counts: if error_filter is None or error not in error_filter: a__: List[str] = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} a__: Optional[Any] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: List[str] = test.split('::' )[0] if test.startswith('tests/models/' ): a__: Dict = test.split('/' )[2] else: a__: Any = None return test def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: int = [(x[0], x[1], get_model(x[2] )) for x in logs] a__: List[Any] = [x for x in logs if x[2] is not None] a__: Optional[Any] = {x[2] for x in logs} a__: Dict = {} for test in tests: a__: Union[str, Any] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) a__: Union[str, Any] = counter.most_common() a__: List[str] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} a__: List[Any] = sum(error_counts.values() ) if n_errors > 0: a__: Any = {'count': n_errors, 'errors': error_counts} a__: Optional[int] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Any = '| no. | error | status |' a__: Any = '|-:|:-|:-|' a__: str = [header, sep] for error in reduced_by_error: a__: int = reduced_by_error[error]['count'] a__: Tuple = F'| {count} | {error[:100]} | |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->str: a__: List[str] = '| model | no. of errors | major error | count |' a__: str = '|-:|-:|-:|-:|' a__: int = [header, sep] for model in reduced_by_model: a__: Tuple = reduced_by_model[model]['count'] a__ , a__: Dict = list(reduced_by_model[model]['errors'].items() )[0] a__: Dict = F'| {model} | {count} | {error[:60]} | {_count} |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') lowercase__ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowercase__ = get_job_links(args.workflow_run_id, token=args.token) lowercase__ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowercase__ = k.find(' / ') lowercase__ = k[index + len(' / ') :] lowercase__ = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowercase__ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowercase__ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowercase__ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowercase__ = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowercase__ = reduce_by_error(errors) lowercase__ = reduce_by_model(errors) lowercase__ = make_github_table(reduced_by_error) lowercase__ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
290
1
"""simple docstring""" import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=4 , ) -> List[Any]: '''simple docstring''' a__: int = parent a__: Optional[int] = batch_size a__: str = seq_length a__: Union[str, Any] = is_training a__: Tuple = use_attention_mask a__: Any = use_token_type_ids a__: Tuple = use_labels a__: int = vocab_size a__: Any = hidden_size a__: Tuple = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Union[str, Any] = intermediate_size a__: Tuple = hidden_act a__: List[str] = hidden_dropout_prob a__: Union[str, Any] = attention_probs_dropout_prob a__: Union[str, Any] = max_position_embeddings a__: Optional[int] = type_vocab_size a__: Tuple = type_sequence_label_size a__: Optional[Any] = initializer_range a__: Any = num_choices def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a__: List[Any] = None if self.use_attention_mask: a__: Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length]) a__: List[Any] = None if self.use_token_type_ids: a__: Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) a__: Tuple = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[int] = self.prepare_config_and_inputs() a__ , a__ , a__ , a__: Optional[int] = config_and_inputs a__: List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = self.prepare_config_and_inputs() a__ , a__ , a__ , a__: List[str] = config_and_inputs a__: Dict = True a__: Optional[Any] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) a__: Any = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = True a__ = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Tuple = FlaxBertModelTester(self) @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Any = FlaxBertModel.from_pretrained('bert-base-cased') a__: str = model(np.ones((1, 1))) self.assertIsNotNone(lowercase)
290
"""simple docstring""" import math def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __a ( _SCREAMING_SNAKE_CASE = 0.1 ) ->int: a__: str = 3 a__: Optional[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_SCREAMING_SNAKE_CASE ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name lowercase__ = '\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)["depth"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline("depth-estimation")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to("cuda")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to("cuda")\n\n\n >>> img = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/cat.png"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda")\n\n >>> prompt = "A robot, 4k photo"\n >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature"\n\n >>> generator = torch.Generator(device="cuda").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save("robot_cat.png")\n ```\n' def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=8 ) ->int: a__: List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 a__: Union[str, Any] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , ) -> Any: '''simple docstring''' super().__init__() self.register_modules( unet=lowercase , scheduler=lowercase , movq=lowercase , ) a__: List[str] = 2 ** (len(self.movq.config.block_out_channels) - 1) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' if latents is None: a__: Optional[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) else: if latents.shape != shape: raise ValueError(f'Unexpected latents shape, got {latents.shape}, expected {shape}') a__: Any = latents.to(lowercase) a__: List[str] = latents * scheduler.init_noise_sigma return latents def lowerCamelCase_ ( self , lowercase=0) -> Union[str, Any]: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`') a__: List[str] = torch.device(f'cuda:{gpu_id}') a__: Any = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase , lowercase) def lowerCamelCase_ ( self , lowercase=0) -> List[Any]: '''simple docstring''' if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0'): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.') a__: Optional[Any] = torch.device(f'cuda:{gpu_id}') if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=lowercase) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) a__: str = None for cpu_offloaded_model in [self.unet, self.movq]: a__ , a__: int = cpu_offload_with_hook(lowercase , lowercase , prev_module_hook=lowercase) # We'll offload the last model manually. a__: Optional[int] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' if not hasattr(self.unet , '_hf_hook'): return self.device for module in self.unet.modules(): if ( hasattr(lowercase , '_hf_hook') and hasattr(module._hf_hook , 'execution_device') and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device) return self.device @torch.no_grad() @replace_example_docstring(lowercase) def __call__( self , lowercase , lowercase , lowercase , lowercase = 5_12 , lowercase = 5_12 , lowercase = 1_00 , lowercase = 4.0 , lowercase = 1 , lowercase = None , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[str, Any]: '''simple docstring''' a__: Tuple = self._execution_device a__: List[Any] = guidance_scale > 1.0 if isinstance(lowercase , lowercase): a__: List[str] = torch.cat(lowercase , dim=0) if isinstance(lowercase , lowercase): a__: Any = torch.cat(lowercase , dim=0) if isinstance(lowercase , lowercase): a__: Any = torch.cat(lowercase , dim=0) a__: List[str] = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: a__: Union[str, Any] = image_embeds.repeat_interleave(lowercase , dim=0) a__: List[str] = negative_image_embeds.repeat_interleave(lowercase , dim=0) a__: List[Any] = hint.repeat_interleave(lowercase , dim=0) a__: Optional[int] = torch.cat([negative_image_embeds, image_embeds] , dim=0).to(dtype=self.unet.dtype , device=lowercase) a__: Any = torch.cat([hint, hint] , dim=0).to(dtype=self.unet.dtype , device=lowercase) self.scheduler.set_timesteps(lowercase , device=lowercase) a__: int = self.scheduler.timesteps a__: Any = self.movq.config.latent_channels a__ , a__: str = downscale_height_and_width(lowercase , lowercase , self.movq_scale_factor) # create initial latent a__: Union[str, Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowercase , lowercase , lowercase , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowercase)): # expand the latents if we are doing classifier free guidance a__: Dict = torch.cat([latents] * 2) if do_classifier_free_guidance else latents a__: Union[str, Any] = {'image_embeds': image_embeds, 'hint': hint} a__: Optional[int] = self.unet( sample=lowercase , timestep=lowercase , encoder_hidden_states=lowercase , added_cond_kwargs=lowercase , return_dict=lowercase , )[0] if do_classifier_free_guidance: a__ , a__: Optional[Any] = noise_pred.split(latents.shape[1] , dim=1) a__ , a__: Dict = noise_pred.chunk(2) a__ , a__: Optional[Any] = variance_pred.chunk(2) a__: Dict = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) a__: Dict = torch.cat([noise_pred, variance_pred_text] , dim=1) if not ( hasattr(self.scheduler.config , 'variance_type') and self.scheduler.config.variance_type in ["learned", "learned_range"] ): a__ , a__: List[Any] = noise_pred.split(latents.shape[1] , dim=1) # compute the previous noisy sample x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , generator=lowercase , )[0] # post-processing a__: Union[str, Any] = self.movq.decode(lowercase , force_not_quantize=lowercase)['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}') if output_type in ["np", "pil"]: a__: List[Any] = image * 0.5 + 0.5 a__: List[Any] = image.clamp(0 , 1) a__: Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1).float().numpy() if output_type == "pil": a__: str = self.numpy_to_pil(lowercase) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase)
290
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ = {'processing_wav2vec2_with_lm': ['Wav2Vec2ProcessorWithLM']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor lowercase__ = logging.get_logger(__name__) class __snake_case ( __lowerCAmelCase ): def __init__( self , *lowercase , **lowercase) -> None: '''simple docstring''' warnings.warn( 'The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use GLPNImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase)
290
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = KandinskyInpaintPipeline a__ = ["""prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] a__ = [ """prompt""", """negative_prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] a__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """negative_prompt""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] a__ = False @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 1_00 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[int] = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base') return tokenizer @property def lowerCamelCase_ ( self) -> Any: '''simple docstring''' torch.manual_seed(0) a__: Dict = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=10_05 , ) a__: Optional[Any] = MultilingualCLIP(lowercase) a__: int = text_encoder.eval() return text_encoder @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' torch.manual_seed(0) a__: Any = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } a__: str = UNetaDConditionModel(**lowercase) return model @property def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' torch.manual_seed(0) a__: Any = VQModel(**self.dummy_movq_kwargs) return model def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = self.dummy_text_encoder a__: int = self.dummy_tokenizer a__: str = self.dummy_unet a__: Any = self.dummy_movq a__: Tuple = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.00085 , beta_end=0.012 , clip_sample=lowercase , set_alpha_to_one=lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=lowercase , ) a__: Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> Any: '''simple docstring''' a__: List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase)).to(lowercase) a__: int = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1)).to(lowercase) # create init_image a__: Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) a__: int = image.cpu().permute(0 , 2 , 3 , 1)[0] a__: Optional[int] = Image.fromarray(np.uinta(lowercase)).convert('RGB').resize((2_56, 2_56)) # create mask a__: Tuple = np.ones((64, 64) , dtype=np.floataa) a__: Optional[Any] = 0 if str(lowercase).startswith('mps'): a__: str = torch.manual_seed(lowercase) else: a__: Dict = torch.Generator(device=lowercase).manual_seed(lowercase) a__: Optional[int] = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[Any] = 'cpu' a__: List[Any] = self.get_dummy_components() a__: Optional[Any] = self.pipeline_class(**lowercase) a__: str = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[int] = pipe(**self.get_dummy_inputs(lowercase)) a__: List[str] = output.images a__: int = pipe( **self.get_dummy_inputs(lowercase) , return_dict=lowercase , )[0] a__: Optional[Any] = image[0, -3:, -3:, -1] a__: List[Any] = image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}') assert image.shape == (1, 64, 64, 3) a__: str = np.array( [0.8326919, 0.73790467, 0.20918581, 0.9309612, 0.5511791, 0.43713328, 0.5513321, 0.49922934, 0.59497786]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy') a__: int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png') a__: Union[str, Any] = np.ones((7_68, 7_68) , dtype=np.floataa) a__: int = 0 a__: Optional[int] = 'a hat' a__: int = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa) pipe_prior.to(lowercase) a__: Any = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa) a__: Optional[Any] = pipeline.to(lowercase) pipeline.set_progress_bar_config(disable=lowercase) a__: Dict = torch.Generator(device='cpu').manual_seed(0) a__ , a__: Optional[Any] = pipe_prior( lowercase , generator=lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() a__: List[str] = pipeline( lowercase , image=lowercase , mask_image=lowercase , image_embeds=lowercase , negative_image_embeds=lowercase , generator=lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) a__: str = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(lowercase , lowercase)
290
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_nllb import NllbTokenizer else: lowercase__ = None lowercase__ = logging.get_logger(__name__) lowercase__ = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': { 'facebook/nllb-200-distilled-600M': ( 'https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/nllb-200-distilled-600M': ( 'https://huggingface.co/facebook/nllb-200-distilled-600M/resolve/main/tokenizer.json' ), }, } lowercase__ = { 'facebook/nllb-large-en-ro': 1024, 'facebook/nllb-200-distilled-600M': 1024, } # fmt: off lowercase__ = ['ace_Arab', 'ace_Latn', 'acm_Arab', 'acq_Arab', 'aeb_Arab', 'afr_Latn', 'ajp_Arab', 'aka_Latn', 'amh_Ethi', 'apc_Arab', 'arb_Arab', 'ars_Arab', 'ary_Arab', 'arz_Arab', 'asm_Beng', 'ast_Latn', 'awa_Deva', 'ayr_Latn', 'azb_Arab', 'azj_Latn', 'bak_Cyrl', 'bam_Latn', 'ban_Latn', 'bel_Cyrl', 'bem_Latn', 'ben_Beng', 'bho_Deva', 'bjn_Arab', 'bjn_Latn', 'bod_Tibt', 'bos_Latn', 'bug_Latn', 'bul_Cyrl', 'cat_Latn', 'ceb_Latn', 'ces_Latn', 'cjk_Latn', 'ckb_Arab', 'crh_Latn', 'cym_Latn', 'dan_Latn', 'deu_Latn', 'dik_Latn', 'dyu_Latn', 'dzo_Tibt', 'ell_Grek', 'eng_Latn', 'epo_Latn', 'est_Latn', 'eus_Latn', 'ewe_Latn', 'fao_Latn', 'pes_Arab', 'fij_Latn', 'fin_Latn', 'fon_Latn', 'fra_Latn', 'fur_Latn', 'fuv_Latn', 'gla_Latn', 'gle_Latn', 'glg_Latn', 'grn_Latn', 'guj_Gujr', 'hat_Latn', 'hau_Latn', 'heb_Hebr', 'hin_Deva', 'hne_Deva', 'hrv_Latn', 'hun_Latn', 'hye_Armn', 'ibo_Latn', 'ilo_Latn', 'ind_Latn', 'isl_Latn', 'ita_Latn', 'jav_Latn', 'jpn_Jpan', 'kab_Latn', 'kac_Latn', 'kam_Latn', 'kan_Knda', 'kas_Arab', 'kas_Deva', 'kat_Geor', 'knc_Arab', 'knc_Latn', 'kaz_Cyrl', 'kbp_Latn', 'kea_Latn', 'khm_Khmr', 'kik_Latn', 'kin_Latn', 'kir_Cyrl', 'kmb_Latn', 'kon_Latn', 'kor_Hang', 'kmr_Latn', 'lao_Laoo', 'lvs_Latn', 'lij_Latn', 'lim_Latn', 'lin_Latn', 'lit_Latn', 'lmo_Latn', 'ltg_Latn', 'ltz_Latn', 'lua_Latn', 'lug_Latn', 'luo_Latn', 'lus_Latn', 'mag_Deva', 'mai_Deva', 'mal_Mlym', 'mar_Deva', 'min_Latn', 'mkd_Cyrl', 'plt_Latn', 'mlt_Latn', 'mni_Beng', 'khk_Cyrl', 'mos_Latn', 'mri_Latn', 'zsm_Latn', 'mya_Mymr', 'nld_Latn', 'nno_Latn', 'nob_Latn', 'npi_Deva', 'nso_Latn', 'nus_Latn', 'nya_Latn', 'oci_Latn', 'gaz_Latn', 'ory_Orya', 'pag_Latn', 'pan_Guru', 'pap_Latn', 'pol_Latn', 'por_Latn', 'prs_Arab', 'pbt_Arab', 'quy_Latn', 'ron_Latn', 'run_Latn', 'rus_Cyrl', 'sag_Latn', 'san_Deva', 'sat_Beng', 'scn_Latn', 'shn_Mymr', 'sin_Sinh', 'slk_Latn', 'slv_Latn', 'smo_Latn', 'sna_Latn', 'snd_Arab', 'som_Latn', 'sot_Latn', 'spa_Latn', 'als_Latn', 'srd_Latn', 'srp_Cyrl', 'ssw_Latn', 'sun_Latn', 'swe_Latn', 'swh_Latn', 'szl_Latn', 'tam_Taml', 'tat_Cyrl', 'tel_Telu', 'tgk_Cyrl', 'tgl_Latn', 'tha_Thai', 'tir_Ethi', 'taq_Latn', 'taq_Tfng', 'tpi_Latn', 'tsn_Latn', 'tso_Latn', 'tuk_Latn', 'tum_Latn', 'tur_Latn', 'twi_Latn', 'tzm_Tfng', 'uig_Arab', 'ukr_Cyrl', 'umb_Latn', 'urd_Arab', 'uzn_Latn', 'vec_Latn', 'vie_Latn', 'war_Latn', 'wol_Latn', 'xho_Latn', 'ydd_Hebr', 'yor_Latn', 'yue_Hant', 'zho_Hans', 'zho_Hant', 'zul_Latn'] class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = ["""input_ids""", """attention_mask"""] a__ = NllbTokenizer a__ = [] a__ = [] def __init__( self , lowercase=None , lowercase=None , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=None , lowercase=None , lowercase=None , lowercase=False , **lowercase , ) -> str: '''simple docstring''' a__: int = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase) if isinstance(lowercase , lowercase) else mask_token a__: Optional[int] = legacy_behaviour super().__init__( vocab_file=lowercase , tokenizer_file=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , src_lang=lowercase , tgt_lang=lowercase , additional_special_tokens=lowercase , legacy_behaviour=lowercase , **lowercase , ) a__: List[Any] = vocab_file a__: int = False if not self.vocab_file else True a__: Any = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) a__: int = { lang_code: self.convert_tokens_to_ids(lowercase) for lang_code in FAIRSEQ_LANGUAGE_CODES } a__: List[Any] = src_lang if src_lang is not None else 'eng_Latn' a__: Union[str, Any] = self.convert_tokens_to_ids(self._src_lang) a__: List[Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Any = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' a__: Tuple = [self.sep_token_id] a__: Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , **lowercase) -> Dict: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') a__: List[str] = src_lang a__: Optional[Any] = self(lowercase , add_special_tokens=lowercase , return_tensors=lowercase , **lowercase) a__: Optional[int] = self.convert_tokens_to_ids(lowercase) a__: str = tgt_lang_id return inputs def lowerCamelCase_ ( self , lowercase , lowercase = "eng_Latn" , lowercase = None , lowercase = "fra_Latn" , **lowercase , ) -> BatchEncoding: '''simple docstring''' a__: str = src_lang a__: str = tgt_lang return super().prepare_seqaseq_batch(lowercase , lowercase , **lowercase) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: int = self.convert_tokens_to_ids(lowercase) if self.legacy_behaviour: a__: Optional[Any] = [] a__: Dict = [self.eos_token_id, self.cur_lang_code] else: a__: str = [self.cur_lang_code] a__: int = [self.eos_token_id] a__: Tuple = self.convert_ids_to_tokens(self.prefix_tokens) a__: Union[str, Any] = self.convert_ids_to_tokens(self.suffix_tokens) a__: str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: int = self.convert_tokens_to_ids(lowercase) if self.legacy_behaviour: a__: int = [] a__: Any = [self.eos_token_id, self.cur_lang_code] else: a__: int = [self.cur_lang_code] a__: int = [self.eos_token_id] a__: Dict = self.convert_ids_to_tokens(self.prefix_tokens) a__: Dict = self.convert_ids_to_tokens(self.suffix_tokens) a__: Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowercase): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return a__: Dict = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase): copyfile(self.vocab_file , lowercase) return (out_vocab_file,)
290
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() lowercase__ = logging.get_logger('transformers.models.encodec') lowercase__ = { 'quantizer.vq.layers.*._codebook.inited': 'quantizer.layers.*.codebook.inited', 'quantizer.vq.layers.*._codebook.cluster_size': 'quantizer.layers.*.codebook.cluster_size', 'quantizer.vq.layers.*._codebook.embed': 'quantizer.layers.*.codebook.embed', 'quantizer.vq.layers.*._codebook.embed_avg': 'quantizer.layers.*.codebook.embed_avg', } lowercase__ = { 'encoder.model.0.conv.conv': 'encoder.layers.0.conv', 'encoder.model.1.block.1.conv.conv': 'encoder.layers.1.block.1.conv', 'encoder.model.1.block.3.conv.conv': 'encoder.layers.1.block.3.conv', 'encoder.model.1.shortcut.conv.conv': 'encoder.layers.1.shortcut.conv', 'encoder.model.3.conv.conv': 'encoder.layers.3.conv', 'encoder.model.4.block.1.conv.conv': 'encoder.layers.4.block.1.conv', 'encoder.model.4.block.3.conv.conv': 'encoder.layers.4.block.3.conv', 'encoder.model.4.shortcut.conv.conv': 'encoder.layers.4.shortcut.conv', 'encoder.model.6.conv.conv': 'encoder.layers.6.conv', 'encoder.model.7.block.1.conv.conv': 'encoder.layers.7.block.1.conv', 'encoder.model.7.block.3.conv.conv': 'encoder.layers.7.block.3.conv', 'encoder.model.7.shortcut.conv.conv': 'encoder.layers.7.shortcut.conv', 'encoder.model.9.conv.conv': 'encoder.layers.9.conv', 'encoder.model.10.block.1.conv.conv': 'encoder.layers.10.block.1.conv', 'encoder.model.10.block.3.conv.conv': 'encoder.layers.10.block.3.conv', 'encoder.model.10.shortcut.conv.conv': 'encoder.layers.10.shortcut.conv', 'encoder.model.12.conv.conv': 'encoder.layers.12.conv', 'encoder.model.13.lstm': 'encoder.layers.13.lstm', 'encoder.model.15.conv.conv': 'encoder.layers.15.conv', } lowercase__ = { 'encoder.model.0.conv.norm': 'encoder.layers.0.norm', 'encoder.model.1.block.1.conv.norm': 'encoder.layers.1.block.1.norm', 'encoder.model.1.block.3.conv.norm': 'encoder.layers.1.block.3.norm', 'encoder.model.1.shortcut.conv.norm': 'encoder.layers.1.shortcut.norm', 'encoder.model.3.conv.norm': 'encoder.layers.3.norm', 'encoder.model.4.block.1.conv.norm': 'encoder.layers.4.block.1.norm', 'encoder.model.4.block.3.conv.norm': 'encoder.layers.4.block.3.norm', 'encoder.model.4.shortcut.conv.norm': 'encoder.layers.4.shortcut.norm', 'encoder.model.6.conv.norm': 'encoder.layers.6.norm', 'encoder.model.7.block.1.conv.norm': 'encoder.layers.7.block.1.norm', 'encoder.model.7.block.3.conv.norm': 'encoder.layers.7.block.3.norm', 'encoder.model.7.shortcut.conv.norm': 'encoder.layers.7.shortcut.norm', 'encoder.model.9.conv.norm': 'encoder.layers.9.norm', 'encoder.model.10.block.1.conv.norm': 'encoder.layers.10.block.1.norm', 'encoder.model.10.block.3.conv.norm': 'encoder.layers.10.block.3.norm', 'encoder.model.10.shortcut.conv.norm': 'encoder.layers.10.shortcut.norm', 'encoder.model.12.conv.norm': 'encoder.layers.12.norm', 'encoder.model.15.conv.norm': 'encoder.layers.15.norm', } lowercase__ = { 'decoder.model.0.conv.conv': 'decoder.layers.0.conv', 'decoder.model.1.lstm': 'decoder.layers.1.lstm', 'decoder.model.3.convtr.convtr': 'decoder.layers.3.conv', 'decoder.model.4.block.1.conv.conv': 'decoder.layers.4.block.1.conv', 'decoder.model.4.block.3.conv.conv': 'decoder.layers.4.block.3.conv', 'decoder.model.4.shortcut.conv.conv': 'decoder.layers.4.shortcut.conv', 'decoder.model.6.convtr.convtr': 'decoder.layers.6.conv', 'decoder.model.7.block.1.conv.conv': 'decoder.layers.7.block.1.conv', 'decoder.model.7.block.3.conv.conv': 'decoder.layers.7.block.3.conv', 'decoder.model.7.shortcut.conv.conv': 'decoder.layers.7.shortcut.conv', 'decoder.model.9.convtr.convtr': 'decoder.layers.9.conv', 'decoder.model.10.block.1.conv.conv': 'decoder.layers.10.block.1.conv', 'decoder.model.10.block.3.conv.conv': 'decoder.layers.10.block.3.conv', 'decoder.model.10.shortcut.conv.conv': 'decoder.layers.10.shortcut.conv', 'decoder.model.12.convtr.convtr': 'decoder.layers.12.conv', 'decoder.model.13.block.1.conv.conv': 'decoder.layers.13.block.1.conv', 'decoder.model.13.block.3.conv.conv': 'decoder.layers.13.block.3.conv', 'decoder.model.13.shortcut.conv.conv': 'decoder.layers.13.shortcut.conv', 'decoder.model.15.conv.conv': 'decoder.layers.15.conv', } lowercase__ = { 'decoder.model.0.conv.norm': 'decoder.layers.0.norm', 'decoder.model.3.convtr.norm': 'decoder.layers.3.norm', 'decoder.model.4.block.1.conv.norm': 'decoder.layers.4.block.1.norm', 'decoder.model.4.block.3.conv.norm': 'decoder.layers.4.block.3.norm', 'decoder.model.4.shortcut.conv.norm': 'decoder.layers.4.shortcut.norm', 'decoder.model.6.convtr.norm': 'decoder.layers.6.norm', 'decoder.model.7.block.1.conv.norm': 'decoder.layers.7.block.1.norm', 'decoder.model.7.block.3.conv.norm': 'decoder.layers.7.block.3.norm', 'decoder.model.7.shortcut.conv.norm': 'decoder.layers.7.shortcut.norm', 'decoder.model.9.convtr.norm': 'decoder.layers.9.norm', 'decoder.model.10.block.1.conv.norm': 'decoder.layers.10.block.1.norm', 'decoder.model.10.block.3.conv.norm': 'decoder.layers.10.block.3.norm', 'decoder.model.10.shortcut.conv.norm': 'decoder.layers.10.shortcut.norm', 'decoder.model.12.convtr.norm': 'decoder.layers.12.norm', 'decoder.model.13.block.1.conv.norm': 'decoder.layers.13.block.1.norm', 'decoder.model.13.block.3.conv.norm': 'decoder.layers.13.block.3.norm', 'decoder.model.13.shortcut.conv.norm': 'decoder.layers.13.shortcut.norm', 'decoder.model.15.conv.norm': 'decoder.layers.15.norm', } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } lowercase__ = [] lowercase__ = [] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: for attribute in key.split('.' ): a__: str = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: str = value elif weight_type == "weight_g": a__: int = value elif weight_type == "weight_v": a__: Tuple = value elif weight_type == "bias": a__: Dict = value elif weight_type == "running_mean": a__: Any = value elif weight_type == "running_var": a__: Tuple = value elif weight_type == "num_batches_tracked": a__: List[str] = value elif weight_type == "weight_ih_l0": a__: List[Any] = value elif weight_type == "weight_hh_l0": a__: List[Any] = value elif weight_type == "bias_ih_l0": a__: List[Any] = value elif weight_type == "bias_hh_l0": a__: List[Any] = value elif weight_type == "weight_ih_l1": a__: int = value elif weight_type == "weight_hh_l1": a__: str = value elif weight_type == "bias_ih_l1": a__: Union[str, Any] = value elif weight_type == "bias_hh_l1": a__: Any = value else: a__: Union[str, Any] = value logger.info(F'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: for key in ignore_keys: if key.endswith('.*' ): if name.startswith(key[:-1] ): return True elif ".*." in key: a__ , a__: Optional[Any] = key.split('.*.' ) if prefix in name and suffix in name: return True elif key in name: return True return False def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[Any] = [] if model_name == "encodec_24khz" or "encodec_32khz": a__: Optional[int] = MAPPING_24K elif model_name == "encodec_48khz": a__: List[Any] = MAPPING_48K else: raise ValueError(F'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): logger.info(F'{name} was ignored' ) continue a__: int = False for key, mapped_key in MAPPING.items(): if "*" in key: a__ , a__: str = key.split('.*.' ) if prefix in name and suffix in name: a__: List[str] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('embed' ) and name.endswith('embed_avg' ): continue a__: List[str] = True if "*" in mapped_key: a__: List[str] = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: str = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: int = 'weight_g' elif "weight_v" in name: a__: Dict = 'weight_v' elif "weight_ih_l0" in name: a__: int = 'weight_ih_l0' elif "weight_hh_l0" in name: a__: Union[str, Any] = 'weight_hh_l0' elif "bias_ih_l0" in name: a__: Optional[Any] = 'bias_ih_l0' elif "bias_hh_l0" in name: a__: Optional[int] = 'bias_hh_l0' elif "weight_ih_l1" in name: a__: Dict = 'weight_ih_l1' elif "weight_hh_l1" in name: a__: Optional[Any] = 'weight_hh_l1' elif "bias_ih_l1" in name: a__: List[str] = 'bias_ih_l1' elif "bias_hh_l1" in name: a__: Optional[Any] = 'bias_hh_l1' elif "bias" in name: a__: List[str] = 'bias' elif "weight" in name: a__: Any = 'weight' elif "running_mean" in name: a__: Dict = 'running_mean' elif "running_var" in name: a__: Dict = 'running_var' elif "num_batches_tracked" in name: a__: Dict = 'num_batches_tracked' else: a__: List[str] = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->int: if config_path is not None: a__: Dict = EncodecConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: a__: Tuple = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": a__: Any = [8, 5, 4, 4] a__: List[str] = [2.2] a__: List[Any] = 64 a__: Dict = 32000 a__: Union[str, Any] = 2048 a__: Union[str, Any] = False a__: Any = False a__: Optional[Any] = False elif model_name == "encodec_48khz": a__: Optional[int] = [8, 5, 4, 2] a__: Union[str, Any] = [3.0, 6.0, 12.0, 24.0] a__: List[str] = 48000 a__: Tuple = 2 a__: Optional[Any] = False a__: Optional[int] = 'time_group_norm' a__: Union[str, Any] = True a__: Dict = 1.0 a__: str = 0.01 else: raise ValueError(F'Unknown model name: {model_name}' ) a__: Optional[int] = EncodecModel(_SCREAMING_SNAKE_CASE ) a__: List[str] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: int = torch.load(_SCREAMING_SNAKE_CASE ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights a__: str = original_checkpoint['best_state'] recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if repo_id: print('Pushing to the hub...' ) feature_extractor.push_to_hub(_SCREAMING_SNAKE_CASE ) model.push_to_hub(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
290
1
"""simple docstring""" import numpy as np def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = int(np.ceil((x_end - xa) / h ) ) a__: str = np.zeros((n + 1,) ) a__: str = ya a__: int = xa for k in range(_SCREAMING_SNAKE_CASE ): a__: int = f(_SCREAMING_SNAKE_CASE , y[k] ) a__: Union[str, Any] = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) a__: int = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) a__: Optional[Any] = f(x + h , y[k] + h * ka ) a__: List[Any] = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: print('moving disk from' , _SCREAMING_SNAKE_CASE , 'to' , _SCREAMING_SNAKE_CASE ) def __a ( ) ->List[str]: a__: Dict = int(input('Height of hanoi: ' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
290
1
"""simple docstring""" import gc import inspect import unittest import torch from parameterized import parameterized from diffusers import PriorTransformer from diffusers.utils import floats_tensor, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = PriorTransformer a__ = """hidden_states""" @property def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Any = 4 a__: Tuple = 8 a__: Dict = 7 a__: List[Any] = floats_tensor((batch_size, embedding_dim)).to(lowercase) a__: List[Any] = floats_tensor((batch_size, embedding_dim)).to(lowercase) a__: List[Any] = floats_tensor((batch_size, num_embeddings, embedding_dim)).to(lowercase) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def lowerCamelCase_ ( self , lowercase=0) -> Any: '''simple docstring''' torch.manual_seed(lowercase) a__: str = 4 a__: Tuple = 8 a__: Union[str, Any] = 7 a__: Union[str, Any] = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: str = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: Optional[int] = torch.randn((batch_size, num_embeddings, embedding_dim)).to(lowercase) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' return (4, 8) @property def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return (4, 8) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[Any] = { 'num_attention_heads': 2, 'attention_head_dim': 4, 'num_layers': 2, 'embedding_dim': 8, 'num_embeddings': 7, 'additional_embeddings': 4, } a__: str = self.dummy_input return init_dict, inputs_dict def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__ , a__: str = PriorTransformer.from_pretrained( 'hf-internal-testing/prior-dummy' , output_loading_info=lowercase) self.assertIsNotNone(lowercase) self.assertEqual(len(loading_info['missing_keys']) , 0) model.to(lowercase) a__: Optional[int] = model(**self.dummy_input)[0] assert hidden_states is not None, "Make sure output is not None" def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__ , a__: Optional[int] = self.prepare_init_args_and_inputs_for_common() a__: Dict = self.model_class(**lowercase) a__: List[str] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__: List[str] = [*signature.parameters.keys()] a__: Tuple = ['hidden_states', 'timestep'] self.assertListEqual(arg_names[:2] , lowercase) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Tuple = PriorTransformer.from_pretrained('hf-internal-testing/prior-dummy') a__: Any = model.to(lowercase) if hasattr(lowercase , 'set_default_attn_processor'): model.set_default_attn_processor() a__: Optional[int] = self.get_dummy_seed_input() with torch.no_grad(): a__: Any = model(**lowercase)[0] a__: int = output[0, :5].flatten().cpu() print(lowercase) # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. a__: Union[str, Any] = torch.tensor([-1.3436, -0.2870, 0.7538, 0.4368, -0.0239]) self.assertTrue(torch_all_close(lowercase , lowercase , rtol=1e-2)) @slow class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self , lowercase=1 , lowercase=7_68 , lowercase=77 , lowercase=0) -> List[str]: '''simple docstring''' torch.manual_seed(lowercase) a__: Optional[Any] = batch_size a__: int = embedding_dim a__: Union[str, Any] = num_embeddings a__: Tuple = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: Tuple = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: Union[str, Any] = torch.randn((batch_size, num_embeddings, embedding_dim)).to(lowercase) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @parameterized.expand( [ # fmt: off [13, [-0.5861, 0.1283, -0.0931, 0.0882, 0.4476, 0.1329, -0.0498, 0.0640]], [37, [-0.4913, 0.0110, -0.0483, 0.0541, 0.4954, -0.0170, 0.0354, 0.1651]], # fmt: on ]) def lowerCamelCase_ ( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = PriorTransformer.from_pretrained('kandinsky-community/kandinsky-2-1-prior' , subfolder='prior') model.to(lowercase) a__: Any = self.get_dummy_seed_input(seed=lowercase) with torch.no_grad(): a__: Tuple = model(**lowercase)[0] assert list(sample.shape) == [1, 7_68] a__: Any = sample[0, :8].flatten().cpu() print(lowercase) a__: Optional[Any] = torch.tensor(lowercase) assert torch_all_close(lowercase , lowercase , atol=1e-3)
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = False ) ->str: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = F'Expected string as input, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[str] = F'Expected boolean as use_pascal parameter, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) a__: int = input_str.split('_' ) a__: List[str] = 0 if use_pascal else 1 a__: List[str] = words[start_index:] a__: List[str] = [word[0].upper() + word[1:] for word in words_to_capitalize] a__: List[str] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
290
1
"""simple docstring""" import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class __snake_case : def __init__( self , lowercase , lowercase=14 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase=None , ) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = parent a__: List[Any] = batch_size a__: Optional[int] = seq_length a__: List[Any] = is_training a__: List[str] = use_token_type_ids a__: Tuple = use_input_mask a__: Union[str, Any] = use_labels a__: Tuple = use_mc_token_ids a__: Any = vocab_size a__: Dict = hidden_size a__: Optional[int] = num_hidden_layers a__: List[Any] = num_attention_heads a__: str = intermediate_size a__: Tuple = hidden_act a__: List[Any] = hidden_dropout_prob a__: Optional[int] = attention_probs_dropout_prob a__: List[str] = max_position_embeddings a__: Tuple = type_vocab_size a__: int = type_sequence_label_size a__: Union[str, Any] = initializer_range a__: Any = num_labels a__: Optional[Any] = num_choices a__: Dict = scope a__: Any = self.vocab_size - 1 def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a__: Tuple = None if self.use_input_mask: a__: Any = random_attention_mask([self.batch_size, self.seq_length]) a__: List[Any] = None if self.use_token_type_ids: a__: Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) a__: Dict = None if self.use_mc_token_ids: a__: int = ids_tensor([self.batch_size, self.num_choices] , self.seq_length) a__: Optional[Any] = None a__: Tuple = None a__: Tuple = None if self.use_labels: a__: List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size) a__: Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) a__: List[str] = ids_tensor([self.batch_size] , self.num_choices) a__: Optional[int] = self.get_config() a__: Any = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase) -> List[str]: '''simple docstring''' a__: List[str] = CTRLModel(config=lowercase) model.to(lowercase) model.eval() model(lowercase , token_type_ids=lowercase , head_mask=lowercase) model(lowercase , token_type_ids=lowercase) a__: str = model(lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(len(result.past_key_values) , config.n_layer) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase) -> int: '''simple docstring''' a__: Optional[Any] = CTRLLMHeadModel(lowercase) model.to(lowercase) model.eval() a__: List[str] = model(lowercase , token_type_ids=lowercase , labels=lowercase) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.prepare_config_and_inputs() ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ): List[Any] = config_and_inputs a__: Optional[Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask} return config, inputs_dict def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , *lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = self.num_labels a__: Optional[int] = CTRLForSequenceClassification(lowercase) model.to(lowercase) model.eval() a__: Any = ids_tensor([self.batch_size] , self.type_sequence_label_size) a__: Optional[Any] = model(lowercase , token_type_ids=lowercase , labels=lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) @require_torch class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () a__ = (CTRLLMHeadModel,) if is_torch_available() else () a__ = ( { """feature-extraction""": CTRLModel, """text-classification""": CTRLForSequenceClassification, """text-generation""": CTRLLMHeadModel, """zero-shot""": CTRLForSequenceClassification, } if is_torch_available() else {} ) a__ = True a__ = False a__ = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase) -> Tuple: '''simple docstring''' if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = CTRLModelTester(self) a__: Optional[Any] = ConfigTester(self , config_class=lowercase , n_embd=37) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*lowercase) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowercase) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def lowerCamelCase_ ( self) -> int: '''simple docstring''' pass @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__: int = CTRLModel.from_pretrained(lowercase) self.assertIsNotNone(lowercase) @unittest.skip('The model doesn\'t support left padding') # and it's not used enough to be worth fixing :) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' pass @require_torch class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Dict = CTRLLMHeadModel.from_pretrained('ctrl') model.to(lowercase) a__: Optional[int] = torch.tensor( [[1_18_59, 0, 16_11, 8]] , dtype=torch.long , device=lowercase) # Legal the president is a__: List[Any] = [ 1_18_59, 0, 16_11, 8, 5, 1_50, 2_64_49, 2, 19, 3_48, 4_69, 3, 25_95, 48, 2_07_40, 24_65_33, 24_65_33, 19, 30, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a a__: List[str] = model.generate(lowercase , do_sample=lowercase) self.assertListEqual(output_ids[0].tolist() , lowercase)
290
"""simple docstring""" class __snake_case : def __init__( self , lowercase , lowercase=None , lowercase=None) -> List[str]: '''simple docstring''' a__: Dict = data a__: List[Any] = previous a__: Any = next_node def __str__( self) -> str: '''simple docstring''' return f'{self.data}' def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.data def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.next def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return self.previous class __snake_case : def __init__( self , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = head def __iter__( self) -> List[Any]: '''simple docstring''' return self def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if not self.current: raise StopIteration else: a__: Dict = self.current.get_data() a__: Optional[Any] = self.current.get_next() return value class __snake_case : def __init__( self) -> Dict: '''simple docstring''' a__: List[Any] = None # First node in list a__: Optional[int] = None # Last node in list def __str__( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.head a__: Optional[Any] = [] while current is not None: nodes.append(current.get_data()) a__: str = current.get_next() return " ".join(str(lowercase) for node in nodes) def __contains__( self , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = self.head while current: if current.get_data() == value: return True a__: Dict = current.get_next() return False def __iter__( self) -> int: '''simple docstring''' return LinkedListIterator(self.head) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: a__: Optional[Any] = node a__: Optional[Any] = node else: self.insert_before_node(self.head , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: self.set_head(lowercase) else: self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = Node(lowercase) if self.head is None: self.set_head(lowercase) else: self.set_tail(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Union[str, Any] = node a__: Optional[Any] = node.previous if node.get_previous() is None: a__: Tuple = node_to_insert else: a__: int = node_to_insert a__: Optional[int] = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Optional[int] = node a__: Tuple = node.next if node.get_next() is None: a__: Optional[int] = node_to_insert else: a__: Any = node_to_insert a__: str = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Any = 1 a__: Tuple = Node(lowercase) a__: Tuple = self.head while node: if current_position == position: self.insert_before_node(lowercase , lowercase) return current_position += 1 a__: List[Any] = node.next self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> Node: '''simple docstring''' a__: Tuple = self.head while node: if node.get_data() == item: return node a__: List[str] = node.get_next() raise Exception('Node not found') def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' if (node := self.get_node(lowercase)) is not None: if node == self.head: a__: Any = self.head.get_next() if node == self.tail: a__: List[Any] = self.tail.get_previous() self.remove_node_pointers(lowercase) @staticmethod def lowerCamelCase_ ( lowercase) -> None: '''simple docstring''' if node.get_next(): a__: Any = node.previous if node.get_previous(): a__: List[str] = node.next a__: int = None a__: Union[str, Any] = None def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self.head is None def __a ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: while a != 0: a__ , a__: List[str] = b % a, a return b def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: if gcd(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) != 1: a__: Dict = F'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Union[str, Any] = 1, 0, a a__ , a__ , a__: Any = 0, 1, m while va != 0: a__: int = ua // va a__ , a__ , a__ , a__ , a__ , a__: Any = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
290
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { 'configuration_blenderbot': [ 'BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlenderbotConfig', 'BlenderbotOnnxConfig', ], 'tokenization_blenderbot': ['BlenderbotTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['BlenderbotTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlenderbotForCausalLM', 'BlenderbotForConditionalGeneration', 'BlenderbotModel', 'BlenderbotPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'TFBlenderbotForConditionalGeneration', 'TFBlenderbotModel', 'TFBlenderbotPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'FlaxBlenderbotForConditionalGeneration', 'FlaxBlenderbotModel', 'FlaxBlenderbotPreTrainedModel', ] if TYPE_CHECKING: from .configuration_blenderbot import ( BLENDERBOT_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotConfig, BlenderbotOnnxConfig, ) from .tokenization_blenderbot import BlenderbotTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_fast import BlenderbotTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot import ( BLENDERBOT_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotForCausalLM, BlenderbotForConditionalGeneration, BlenderbotModel, BlenderbotPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot import ( TFBlenderbotForConditionalGeneration, TFBlenderbotModel, TFBlenderbotPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot import ( FlaxBlenderbotForConditionalGeneration, FlaxBlenderbotModel, FlaxBlenderbotPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowercase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Any: if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): a__: Optional[int] = [image] a__: str = [trans(img.convert('RGB' ) ) for img in image] a__: Any = torch.stack(_SCREAMING_SNAKE_CASE ) return image class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM a__: Dict = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowercase , scheduler=lowercase) def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f'The value of strength should in [0.0, 1.0] but is {strength}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: int = min(int(num_inference_steps * strength) , lowercase) a__: Any = max(num_inference_steps - init_timestep , 0) a__: Union[str, Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> List[Any]: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( f'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase)}') a__: Tuple = image.to(device=lowercase , dtype=lowercase) if isinstance(lowercase , lowercase) and len(lowercase) != batch_size: raise ValueError( f'You have passed a list of generators of length {len(lowercase)}, but requested an effective batch' f' size of {batch_size}. Make sure the batch size matches the length of the generators.') a__: List[str] = init_latents.shape a__: List[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) # get latents print('add noise to latents at timestep' , lowercase) a__: int = self.scheduler.add_noise(lowercase , lowercase , lowercase) a__: Dict = init_latents return latents @torch.no_grad() def __call__( self , lowercase = None , lowercase = 0.8 , lowercase = 1 , lowercase = None , lowercase = 0.0 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase) # 2. Preprocess image a__: Tuple = preprocess(lowercase) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device) a__ , a__: Union[str, Any] = self.get_timesteps(lowercase , lowercase , self.device) a__: Optional[int] = timesteps[:1].repeat(lowercase) # 4. Prepare latent variables a__: Union[str, Any] = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase) a__: Optional[Any] = latents # 5. Denoising loop for t in self.progress_bar(lowercase): # 1. predict noise model_output a__: Dict = self.unet(lowercase , lowercase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample a__: Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1) a__: Optional[int] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": a__: Dict = self.numpy_to_pil(lowercase) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase)
290
1
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowercase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Any: if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): a__: Optional[int] = [image] a__: str = [trans(img.convert('RGB' ) ) for img in image] a__: Any = torch.stack(_SCREAMING_SNAKE_CASE ) return image class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM a__: Dict = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowercase , scheduler=lowercase) def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f'The value of strength should in [0.0, 1.0] but is {strength}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: int = min(int(num_inference_steps * strength) , lowercase) a__: Any = max(num_inference_steps - init_timestep , 0) a__: Union[str, Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> List[Any]: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( f'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase)}') a__: Tuple = image.to(device=lowercase , dtype=lowercase) if isinstance(lowercase , lowercase) and len(lowercase) != batch_size: raise ValueError( f'You have passed a list of generators of length {len(lowercase)}, but requested an effective batch' f' size of {batch_size}. Make sure the batch size matches the length of the generators.') a__: List[str] = init_latents.shape a__: List[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) # get latents print('add noise to latents at timestep' , lowercase) a__: int = self.scheduler.add_noise(lowercase , lowercase , lowercase) a__: Dict = init_latents return latents @torch.no_grad() def __call__( self , lowercase = None , lowercase = 0.8 , lowercase = 1 , lowercase = None , lowercase = 0.0 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase) # 2. Preprocess image a__: Tuple = preprocess(lowercase) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device) a__ , a__: Union[str, Any] = self.get_timesteps(lowercase , lowercase , self.device) a__: Optional[int] = timesteps[:1].repeat(lowercase) # 4. Prepare latent variables a__: Union[str, Any] = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase) a__: Optional[Any] = latents # 5. Denoising loop for t in self.progress_bar(lowercase): # 1. predict noise model_output a__: Dict = self.unet(lowercase , lowercase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample a__: Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1) a__: Optional[int] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": a__: Dict = self.numpy_to_pil(lowercase) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase)
290
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: Optional[int] = SamImageProcessor() a__: Tuple = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> List[Any]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[Any] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Optional[int] = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: List[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = self.get_image_processor() a__: List[Any] = SamProcessor(image_processor=lowercase) a__: Optional[int] = self.prepare_image_inputs() a__: Optional[Any] = image_processor(lowercase , return_tensors='np') a__: Tuple = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_torch def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.get_image_processor() a__: List[str] = SamProcessor(image_processor=lowercase) a__: Optional[Any] = [torch.ones((1, 3, 5, 5))] a__: Union[str, Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: int = processor.post_process_masks(lowercase , lowercase , lowercase) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Optional[int] = processor.post_process_masks( lowercase , torch.tensor(lowercase) , torch.tensor(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Dict = [np.ones((1, 3, 5, 5))] a__: Tuple = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = [[1, 0], [0, 1]] with self.assertRaises(lowercase): a__: List[Any] = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) @require_vision @require_tf class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: List[Any] = SamImageProcessor() a__: Optional[int] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> int: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[int] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Dict = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = self.get_image_processor() a__: str = SamProcessor(image_processor=lowercase) a__: int = self.prepare_image_inputs() a__: int = image_processor(lowercase , return_tensors='np') a__: Dict = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_tf def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Any = SamProcessor(image_processor=lowercase) a__: str = [tf.ones((1, 3, 5, 5))] a__: List[Any] = [[17_64, 26_46]] a__: List[Any] = [[6_83, 10_24]] a__: List[Any] = processor.post_process_masks(lowercase , lowercase , lowercase , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = processor.post_process_masks( lowercase , tf.convert_to_tensor(lowercase) , tf.convert_to_tensor(lowercase) , return_tensors='tf' , ) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Optional[Any] = [np.ones((1, 3, 5, 5))] a__: int = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: List[str] = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError): a__: Any = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = tempfile.mkdtemp() a__: int = SamImageProcessor() a__: Union[str, Any] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Any = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.get_image_processor() a__: int = SamProcessor(image_processor=lowercase) a__: int = np.random.randint(0 , 2 , size=(1, 3, 5, 5)).astype(np.floataa) a__: Dict = [tf.convert_to_tensor(lowercase)] a__: Union[str, Any] = [torch.tensor(lowercase)] a__: List[Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: Tuple = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='tf') a__: str = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='pt') self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy())) @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Dict = SamProcessor(image_processor=lowercase) a__: Any = self.prepare_image_inputs() a__: List[Any] = image_processor(lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Tuple = processor(images=lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Any = image_processor(lowercase , return_tensors='tf')['pixel_values'].numpy() a__: Any = processor(images=lowercase , return_tensors='tf')['pixel_values'].numpy() self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase))
290
1
"""simple docstring""" import numpy as np # Importing the Keras libraries and packages import tensorflow as tf from tensorflow.keras import layers, models if __name__ == "__main__": # Initialising the CNN # (Sequential- Building the model layer by layer) lowercase__ = models.Sequential() # Step 1 - Convolution # Here 64,64 is the length & breadth of dataset images and 3 is for the RGB channel # (3,3) is the kernel size (filter matrix) classifier.add( layers.ConvaD(32, (3, 3), input_shape=(64, 64, 3), activation='relu') ) # Step 2 - Pooling classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Adding a second convolutional layer classifier.add(layers.ConvaD(32, (3, 3), activation='relu')) classifier.add(layers.MaxPoolingaD(pool_size=(2, 2))) # Step 3 - Flattening classifier.add(layers.Flatten()) # Step 4 - Full connection classifier.add(layers.Dense(units=128, activation='relu')) classifier.add(layers.Dense(units=1, activation='sigmoid')) # Compiling the CNN classifier.compile( optimizer='adam', loss='binary_crossentropy', metrics=['accuracy'] ) # Part 2 - Fitting the CNN to the images # Load Trained model weights # from keras.models import load_model # regressor=load_model('cnn.h5') lowercase__ = tf.keras.preprocessing.image.ImageDataGenerator( rescale=1.0 / 255, shear_range=0.2, zoom_range=0.2, horizontal_flip=True ) lowercase__ = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1.0 / 255) lowercase__ = train_datagen.flow_from_directory( 'dataset/training_set', target_size=(64, 64), batch_size=32, class_mode='binary' ) lowercase__ = test_datagen.flow_from_directory( 'dataset/test_set', target_size=(64, 64), batch_size=32, class_mode='binary' ) classifier.fit_generator( training_set, steps_per_epoch=5, epochs=30, validation_data=test_set ) classifier.save('cnn.h5') # Part 3 - Making new predictions lowercase__ = tf.keras.preprocessing.image.load_img( 'dataset/single_prediction/image.png', target_size=(64, 64) ) lowercase__ = tf.keras.preprocessing.image.img_to_array(test_image) lowercase__ = np.expand_dims(test_image, axis=0) lowercase__ = classifier.predict(test_image) # training_set.class_indices if result[0][0] == 0: lowercase__ = 'Normal' if result[0][0] == 1: lowercase__ = 'Abnormality detected'
290
"""simple docstring""" from math import pow, sqrt def __a ( *_SCREAMING_SNAKE_CASE ) ->bool: a__: Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) > 0 and all(value > 0.0 for value in values ) return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
1
"""simple docstring""" class __snake_case ( __lowerCAmelCase ): pass class __snake_case ( __lowerCAmelCase ): pass class __snake_case : def __init__( self) -> Optional[Any]: '''simple docstring''' a__: Tuple = [ [], [], [], ] def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' try: if len(self.queues[priority]) >= 1_00: raise OverflowError('Maximum queue size is 100') self.queues[priority].append(lowercase) except IndexError: raise ValueError('Valid priorities are 0, 1, and 2') def lowerCamelCase_ ( self) -> int: '''simple docstring''' for queue in self.queues: if queue: return queue.pop(0) raise UnderFlowError('All queues are empty') def __str__( self) -> str: '''simple docstring''' return "\n".join(f'Priority {i}: {q}' for i, q in enumerate(self.queues)) class __snake_case : def __init__( self) -> int: '''simple docstring''' a__: Optional[int] = [] def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if len(self.queue) == 1_00: raise OverFlowError('Maximum queue size is 100') self.queue.append(lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' if not self.queue: raise UnderFlowError('The queue is empty') else: a__: int = min(self.queue) self.queue.remove(lowercase) return data def __str__( self) -> str: '''simple docstring''' return str(self.queue) def __a ( ) ->int: a__: List[str] = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def __a ( ) ->List[str]: a__: Tuple = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
290
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """roberta-prelayernorm""" def __init__( self , lowercase=5_02_65 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=True , lowercase=None , **lowercase , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase) a__: Union[str, Any] = vocab_size a__: str = hidden_size a__: Tuple = num_hidden_layers a__: List[str] = num_attention_heads a__: Dict = hidden_act a__: int = intermediate_size a__: Tuple = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: Tuple = max_position_embeddings a__: Tuple = type_vocab_size a__: Optional[Any] = initializer_range a__: Tuple = layer_norm_eps a__: Optional[int] = position_embedding_type a__: Any = use_cache a__: Dict = classifier_dropout class __snake_case ( __lowerCAmelCase ): @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": a__: str = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a__: Union[str, Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
290
1
"""simple docstring""" lowercase__ = 'Input must be a string of 8 numbers plus letter' lowercase__ = 'TRWAGMYFPDXBNJZSQVHLCKE' def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: str = F'Expected string as input, found {type(_SCREAMING_SNAKE_CASE ).__name__}' raise TypeError(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = spanish_id.replace('-' , '' ).upper() if len(_SCREAMING_SNAKE_CASE ) != 9: raise ValueError(_SCREAMING_SNAKE_CASE ) try: a__: List[str] = int(spanish_id_clean[0:8] ) a__: int = spanish_id_clean[8] except ValueError as ex: raise ValueError(_SCREAMING_SNAKE_CASE ) from ex if letter.isdigit(): raise ValueError(_SCREAMING_SNAKE_CASE ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """audio-spectrogram-transformer""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=16 , lowercase=True , lowercase=10 , lowercase=10 , lowercase=10_24 , lowercase=1_28 , **lowercase , ) -> str: '''simple docstring''' super().__init__(**lowercase) a__: Any = hidden_size a__: int = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Any = intermediate_size a__: Union[str, Any] = hidden_act a__: int = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: str = initializer_range a__: Tuple = layer_norm_eps a__: Any = patch_size a__: int = qkv_bias a__: Optional[Any] = frequency_stride a__: int = time_stride a__: List[str] = max_length a__: Tuple = num_mel_bins
290
1
"""simple docstring""" import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __snake_case ( __lowerCAmelCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(lowercase , 'hidden_sizes')) self.parent.assertTrue(hasattr(lowercase , 'neck_hidden_sizes')) self.parent.assertTrue(hasattr(lowercase , 'num_attention_heads')) class __snake_case : def __init__( self , lowercase , lowercase=13 , lowercase=32 , lowercase=2 , lowercase=3 , lowercase=6_40 , lowercase=4 , lowercase="silu" , lowercase=3 , lowercase=32 , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=10 , lowercase=None , ) -> List[str]: '''simple docstring''' a__: Union[str, Any] = parent a__: Any = batch_size a__: int = image_size a__: Any = patch_size a__: str = num_channels a__: Union[str, Any] = last_hidden_size a__: int = num_attention_heads a__: Any = hidden_act a__: Union[str, Any] = conv_kernel_size a__: Optional[int] = output_stride a__: str = hidden_dropout_prob a__: Union[str, Any] = attention_probs_dropout_prob a__: List[str] = classifier_dropout_prob a__: Any = use_labels a__: Any = is_training a__: Tuple = num_labels a__: Tuple = initializer_range a__: Optional[Any] = scope def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) a__: Optional[Any] = None a__: Optional[Any] = None if self.use_labels: a__: Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels) a__: List[str] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels) a__: Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: List[str] = MobileViTModel(config=lowercase) model.to(lowercase) model.eval() a__: int = model(lowercase) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> Tuple: '''simple docstring''' a__: Optional[Any] = self.num_labels a__: Union[str, Any] = MobileViTForImageClassification(lowercase) model.to(lowercase) model.eval() a__: str = model(lowercase , labels=lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.num_labels a__: Any = MobileViTForSemanticSegmentation(lowercase) model.to(lowercase) model.eval() a__: Any = model(lowercase) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) a__: int = model(lowercase , labels=lowercase) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Dict = self.prepare_config_and_inputs() a__ , a__ , a__ , a__: Optional[int] = config_and_inputs a__: Union[str, Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) a__ = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) a__ = False a__ = False a__ = False a__ = False def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Optional[Any] = MobileViTModelTester(self) a__: Union[str, Any] = MobileViTConfigTester(self , config_class=lowercase , has_text_modality=lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds') def lowerCamelCase_ ( self) -> str: '''simple docstring''' pass @unittest.skip(reason='MobileViT does not support input and output embeddings') def lowerCamelCase_ ( self) -> str: '''simple docstring''' pass @unittest.skip(reason='MobileViT does not output attentions') def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' pass def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__ , a__: Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__: Tuple = model_class(lowercase) a__: List[Any] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__: List[Any] = [*signature.parameters.keys()] a__: Optional[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowercase) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' pass def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' def check_hidden_states_output(lowercase , lowercase , lowercase): a__: Tuple = model_class(lowercase) model.to(lowercase) model.eval() with torch.no_grad(): a__: Dict = model(**self._prepare_for_class(lowercase , lowercase)) a__: Dict = outputs.hidden_states a__: Union[str, Any] = 5 self.assertEqual(len(lowercase) , lowercase) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. a__: Union[str, Any] = 2 for i in range(len(lowercase)): self.assertListEqual( list(hidden_states[i].shape[-2:]) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2) a__ , a__: Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__: Optional[Any] = True check_hidden_states_output(lowercase , lowercase , lowercase) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a__: Optional[Any] = True check_hidden_states_output(lowercase , lowercase , lowercase) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowercase) @slow def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__: Tuple = MobileViTModel.from_pretrained(lowercase) self.assertIsNotNone(lowercase) def __a ( ) ->Optional[Any]: a__: Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class __snake_case ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small') if is_vision_available() else None @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Tuple = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small').to(lowercase) a__: str = self.default_image_processor a__: Optional[Any] = prepare_img() a__: Optional[int] = image_processor(images=lowercase , return_tensors='pt').to(lowercase) # forward pass with torch.no_grad(): a__: int = model(**lowercase) # verify the logits a__: Optional[int] = torch.Size((1, 10_00)) self.assertEqual(outputs.logits.shape , lowercase) a__: Optional[Any] = torch.tensor([-1.9364, -1.2327, -0.4653]).to(lowercase) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1e-4)) @slow def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: str = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small') a__: Union[str, Any] = model.to(lowercase) a__: str = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small') a__: Any = prepare_img() a__: Optional[Any] = image_processor(images=lowercase , return_tensors='pt').to(lowercase) # forward pass with torch.no_grad(): a__: Union[str, Any] = model(**lowercase) a__: int = outputs.logits # verify the logits a__: int = torch.Size((1, 21, 32, 32)) self.assertEqual(logits.shape , lowercase) a__: Tuple = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=lowercase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowercase , atol=1e-4)) @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Tuple = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small') a__: Tuple = model.to(lowercase) a__: str = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small') a__: Union[str, Any] = prepare_img() a__: Optional[Any] = image_processor(images=lowercase , return_tensors='pt').to(lowercase) # forward pass with torch.no_grad(): a__: Any = model(**lowercase) a__: str = outputs.logits.detach().cpu() a__: Optional[Any] = image_processor.post_process_semantic_segmentation(outputs=lowercase , target_sizes=[(50, 60)]) a__: List[Any] = torch.Size((50, 60)) self.assertEqual(segmentation[0].shape , lowercase) a__: str = image_processor.post_process_semantic_segmentation(outputs=lowercase) a__: Any = torch.Size((32, 32)) self.assertEqual(segmentation[0].shape , lowercase)
290
"""simple docstring""" import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowercase__ = get_tests_dir('fixtures/test_sentencepiece.model') lowercase__ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') lowercase__ = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = CamembertTokenizer a__ = CamembertTokenizerFast a__ = True a__ = True def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing a__: Tuple = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Optional[Any] = '<pad>' a__: List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase) , lowercase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase) , lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: str = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>NOTUSED') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(lowercase) , 10_04) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_05) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[Any] = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) a__: List[Any] = CamembertTokenizerFast.from_pretrained(self.tmpdirname) a__: Dict = 'I was born in 92000, and this is falsé.' a__: Optional[int] = tokenizer.encode(lowercase) a__: Any = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) a__: Optional[Any] = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) a__: Tuple = tokenizer.convert_ids_to_tokens(lowercase) a__: Tuple = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' if not self.test_rust_tokenizer: return a__: Dict = self.get_tokenizer() a__: str = self.get_rust_tokenizer() a__: int = 'I was born in 92000, and this is falsé.' a__: Optional[Any] = tokenizer.tokenize(lowercase) a__: List[Any] = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) a__: str = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) a__: Tuple = self.get_rust_tokenizer() a__: Union[str, Any] = tokenizer.encode(lowercase) a__: List[Any] = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Union[str, Any] = {'input_ids': [[5, 54, 71_96, 2_97, 30, 23, 7_76, 18, 11, 32_15, 37_05, 82_52, 22, 31_64, 11_81, 21_16, 29, 16, 8_13, 25, 7_91, 33_14, 20, 34_46, 38, 2_75_75, 1_20, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 4_68, 17, 11, 90_88, 20, 15_17, 8, 2_28_04, 1_88_18, 10, 38, 6_29, 6_07, 6_07, 1_42, 19, 71_96, 8_67, 56, 1_03_26, 24, 22_67, 20, 4_16, 50_72, 1_56_12, 2_33, 7_34, 7, 23_99, 27, 16, 30_15, 16_49, 7, 24, 20, 43_38, 23_99, 27, 13, 34_00, 14, 13, 61_89, 8, 9_30, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. a__: int = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=lowercase , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=lowercase , )
290
1
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class __snake_case : def __init__( self , lowercase , lowercase=99 , lowercase=13 , lowercase=16 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=2 , lowercase=32 , lowercase=4 , lowercase=4 , lowercase=30 , lowercase=0 , lowercase=1 , lowercase=2 , lowercase=None , ) -> Optional[Any]: '''simple docstring''' a__: Any = parent a__: Union[str, Any] = batch_size a__: str = decoder_seq_length # For common tests a__: Optional[Any] = self.decoder_seq_length a__: Optional[Any] = is_training a__: str = use_attention_mask a__: Optional[int] = use_labels a__: List[Any] = vocab_size a__: Optional[int] = d_model a__: str = d_model a__: int = decoder_layers a__: Dict = decoder_layers a__: str = decoder_ffn_dim a__: Dict = decoder_attention_heads a__: Union[str, Any] = decoder_attention_heads a__: int = eos_token_id a__: Tuple = bos_token_id a__: Optional[Any] = pad_token_id a__: int = decoder_start_token_id a__: Union[str, Any] = use_cache a__: Dict = max_position_embeddings a__: Union[str, Any] = None a__: List[str] = decoder_seq_length a__: Tuple = 2 a__: List[str] = 1 def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Any = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size) a__: str = None if self.use_attention_mask: a__: List[str] = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2) a__: Any = None if self.use_labels: a__: Any = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size) a__: Optional[int] = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , ) -> str: '''simple docstring''' a__: int = True a__: Optional[int] = TrOCRDecoder(config=lowercase).to(lowercase).eval() a__: Union[str, Any] = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass a__: List[Any] = model(lowercase , use_cache=lowercase) a__: int = model(lowercase) a__: List[str] = model(lowercase , use_cache=lowercase) self.parent.assertTrue(len(lowercase) == len(lowercase)) self.parent.assertTrue(len(lowercase) == len(lowercase) + 1) a__: Any = outputs['past_key_values'] # create hypothetical next token and extent to next_input_ids a__: int = ids_tensor((2, 1) , config.vocab_size - 1) + 1 # append to next input_ids and a__: int = torch.cat([input_ids, next_tokens] , dim=-1) a__: Tuple = model(lowercase)['last_hidden_state'] a__: Optional[Any] = model(lowercase , past_key_values=lowercase)['last_hidden_state'] # select random slice a__: Dict = ids_tensor((1,) , output_from_past.shape[-1]).item() a__: Union[str, Any] = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() a__: Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowercase , lowercase , atol=1e-3) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: List[str] = self.prepare_config_and_inputs() a__ , a__ , a__ , a__: List[str] = config_and_inputs a__: Optional[Any] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_torch class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () a__ = (TrOCRForCausalLM,) if is_torch_available() else () a__ = {"""text-generation""": TrOCRForCausalLM} if is_torch_available() else {} a__ = True a__ = False def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = TrOCRStandaloneDecoderModelTester(self , is_training=lowercase) a__: int = ConfigTester(self , config_class=lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' pass def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' pass def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' pass def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowercase) def lowerCamelCase_ ( self) -> str: '''simple docstring''' return @unittest.skip('The model doesn\'t support left padding') # and it's not used enough to be worth fixing :) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' pass
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE = 1000000 ) ->int: a__: int = limit + 1 a__: Optional[int] = [0] * limit for first_term in range(1 , _SCREAMING_SNAKE_CASE ): for n in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a a__: Any = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f"{solution() = }")
290
1
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) lowercase__ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } lowercase__ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: for attribute in key.split('.' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models a__: str = 'lm_head' a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: Tuple = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: Any = value elif weight_type == "weight_g": a__: Optional[int] = value elif weight_type == "weight_v": a__: str = value elif weight_type == "bias": a__: Optional[Any] = value else: a__: List[Any] = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: str = [] a__: Dict = fairseq_model.state_dict() a__: Any = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): a__: Tuple = False if "conv_layers" in name: load_conv_layer( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == 'group' , ) a__: List[str] = True else: for key, mapped_key in MAPPING.items(): a__: Tuple = 'unispeech.' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: a__: Optional[Any] = True if "*" in mapped_key: a__: List[Any] = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: Dict = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: Tuple = 'weight_g' elif "weight_v" in name: a__: str = 'weight_v' elif "bias" in name: a__: Optional[int] = 'bias' elif "weight" in name: # TODO: don't match quantizer.weight_proj a__: List[Any] = 'weight' else: a__: List[Any] = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: a__: str = full_name.split('conv_layers.' )[-1] a__: List[Any] = name.split('.' ) a__: List[Any] = int(items[0] ) a__: Tuple = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) a__: int = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) a__: Dict = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) a__: Optional[int] = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) a__: Union[str, Any] = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(_SCREAMING_SNAKE_CASE ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=True ) ->Union[str, Any]: if config_path is not None: a__: Optional[int] = UniSpeechConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: a__: List[str] = UniSpeechConfig() if is_finetuned: if dict_path: a__: Tuple = Dictionary.load_from_json(_SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq a__: List[str] = target_dict.pad_index a__: Dict = target_dict.bos_index a__: List[str] = target_dict.eos_index a__: Any = len(target_dict.symbols ) a__: Optional[int] = os.path.join(_SCREAMING_SNAKE_CASE , 'vocab.json' ) if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(_SCREAMING_SNAKE_CASE ) ) return os.makedirs(_SCREAMING_SNAKE_CASE , exist_ok=_SCREAMING_SNAKE_CASE ) a__: Optional[int] = target_dict.indices # fairseq has the <pad> and <s> switched a__: List[Any] = 42 a__: Optional[Any] = 43 with open(_SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Dict = WavaVecaPhonemeCTCTokenizer( _SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=_SCREAMING_SNAKE_CASE , ) a__: Union[str, Any] = True if config.feat_extract_norm == 'layer' else False a__: str = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE , ) a__: List[Any] = WavaVecaProcessor(feature_extractor=_SCREAMING_SNAKE_CASE , tokenizer=_SCREAMING_SNAKE_CASE ) processor.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: Dict = UniSpeechForCTC(_SCREAMING_SNAKE_CASE ) else: a__: List[Any] = UniSpeechForPreTraining(_SCREAMING_SNAKE_CASE ) if is_finetuned: a__ , a__ , a__: List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path} ) else: a__ , a__ , a__: List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) a__: str = model[0].eval() recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) hf_unispeech.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) lowercase__ = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
290
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union lowercase__ = TypeVar('T') lowercase__ = Union[List[T], Tuple[T, ...]] lowercase__ = Union[T, List[T], Dict[str, T]] lowercase__ = Union[str, bytes, os.PathLike]
290
1
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
290
"""simple docstring""" from math import pi, sqrt, tan def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) a__: List[Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(_SCREAMING_SNAKE_CASE , 2 ) * torus_radius * tube_radius def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) a__: int = (sidea + sidea + sidea) / 2 a__: Tuple = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(f"Rectangle: {area_rectangle(10, 20) = }") print(f"Square: {area_square(10) = }") print(f"Triangle: {area_triangle(10, 10) = }") print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(f"Parallelogram: {area_parallelogram(10, 20) = }") print(f"Rhombus: {area_rhombus(10, 20) = }") print(f"Trapezium: {area_trapezium(10, 20, 30) = }") print(f"Circle: {area_circle(20) = }") print(f"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(f"Cube: {surface_area_cube(20) = }") print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(f"Sphere: {surface_area_sphere(20) = }") print(f"Hemisphere: {surface_area_hemisphere(20) = }") print(f"Cone: {surface_area_cone(10, 20) = }") print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(f"Cylinder: {surface_area_cylinder(10, 20) = }") print(f"Torus: {surface_area_torus(20, 10) = }") print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(f"Square: {area_reg_polygon(4, 10) = }") print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
290
1
"""simple docstring""" import baseaa def __a ( _SCREAMING_SNAKE_CASE ) ->bytes: return baseaa.aaaencode(string.encode('utf-8' ) ) def __a ( _SCREAMING_SNAKE_CASE ) ->str: return baseaa.aaadecode(_SCREAMING_SNAKE_CASE ).decode('utf-8' ) if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase__ = random.Random() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: if rng is None: a__: Any = global_rng a__: int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=4_00 , lowercase=20_00 , lowercase=1 , lowercase=0.0 , lowercase=1_60_00 , lowercase=True , lowercase=True , ) -> Union[str, Any]: '''simple docstring''' a__: Tuple = parent a__: Optional[int] = batch_size a__: Optional[Any] = min_seq_length a__: Optional[int] = max_seq_length a__: Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a__: Dict = feature_size a__: Any = padding_value a__: Optional[Any] = sampling_rate a__: Optional[Any] = return_attention_mask a__: str = do_normalize def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def lowerCamelCase_ ( self , lowercase=False , lowercase=False) -> Tuple: '''simple docstring''' def _flatten(lowercase): return list(itertools.chain(*lowercase)) if equal_length: a__: Dict = floats_list((self.batch_size, self.max_seq_length)) else: # make sure that inputs increase in size a__: List[Any] = [ _flatten(floats_list((x, self.feature_size))) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: a__: str = [np.asarray(lowercase) for x in speech_inputs] return speech_inputs class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = WavaVecaFeatureExtractor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = WavaVecaFeatureExtractionTester(self) def lowerCamelCase_ ( self , lowercase) -> List[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(lowercase , axis=0) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0) - 1) < 1e-3)) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 a__: Optional[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: List[str] = [np.asarray(lowercase) for speech_input in speech_inputs] # Test not batched input a__: Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='np').input_values a__: Dict = feat_extract(np_speech_inputs[0] , return_tensors='np').input_values self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test batched a__: Dict = feat_extract(lowercase , return_tensors='np').input_values a__: int = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test 2-D numpy arrays are batched. a__: int = [floats_list((1, x))[0] for x in (8_00, 8_00, 8_00)] a__: Union[str, Any] = np.asarray(lowercase) a__: int = feat_extract(lowercase , return_tensors='np').input_values a__: Any = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: List[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Optional[int] = ['longest', 'max_length', 'do_not_pad'] a__: List[Any] = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: Dict = feat_extract(lowercase , padding=lowercase , max_length=lowercase , return_tensors='np') a__: Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self.assertTrue(input_values[0][8_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self.assertTrue(input_values[0][10_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Optional[int] = range(8_00 , 14_00 , 2_00) a__: List[str] = [floats_list((1, x))[0] for x in lengths] a__: Tuple = ['longest', 'max_length', 'do_not_pad'] a__: Dict = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: int = feat_extract(lowercase , max_length=lowercase , padding=lowercase) a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Any = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Dict = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='max_length' , return_tensors='np') a__: int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1]) self._check_zero_mean_unit_variance(input_values[2]) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: int = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: str = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='longest' , return_tensors='np') a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00)) a__: Dict = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Tuple = feat_extract( lowercase , truncation=lowercase , max_length=20_00 , padding='longest' , return_tensors='np') a__: str = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00)) @require_torch def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' import torch a__: Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Tuple = np.random.rand(1_00).astype(np.floataa) a__: Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a__: Any = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np') self.assertTrue(np_processed.input_values.dtype == np.floataa) a__: Optional[Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt') self.assertTrue(pt_processed.input_values.dtype == torch.floataa) @slow @require_torch def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: a__: str = WavaVecaConfig.from_pretrained(lowercase) a__: str = WavaVecaFeatureExtractor.from_pretrained(lowercase) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer')
290
1
"""simple docstring""" from ...processing_utils import ProcessorMixin class __snake_case ( __lowerCAmelCase ): a__ = """SpeechT5FeatureExtractor""" a__ = """SpeechT5Tokenizer""" def __init__( self , lowercase , lowercase) -> Dict: '''simple docstring''' super().__init__(lowercase , lowercase) def __call__( self , *lowercase , **lowercase) -> Tuple: '''simple docstring''' a__: str = kwargs.pop('audio' , lowercase) a__: Dict = kwargs.pop('text' , lowercase) a__: Tuple = kwargs.pop('text_target' , lowercase) a__: Union[str, Any] = kwargs.pop('audio_target' , lowercase) a__: List[str] = kwargs.pop('sampling_rate' , lowercase) if audio is not None and text is not None: raise ValueError( 'Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?') if audio_target is not None and text_target is not None: raise ValueError( 'Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?') if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( 'You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.') if audio is not None: a__: int = self.feature_extractor(lowercase , *lowercase , sampling_rate=lowercase , **lowercase) elif text is not None: a__: str = self.tokenizer(lowercase , **lowercase) else: a__: int = None if audio_target is not None: a__: Tuple = self.feature_extractor(audio_target=lowercase , *lowercase , sampling_rate=lowercase , **lowercase) a__: Any = targets['input_values'] elif text_target is not None: a__: str = self.tokenizer(lowercase , **lowercase) a__: Tuple = targets['input_ids'] else: a__: Tuple = None if inputs is None: return targets if targets is not None: a__: Dict = labels a__: Optional[int] = targets.get('attention_mask') if decoder_attention_mask is not None: a__: List[str] = decoder_attention_mask return inputs def lowerCamelCase_ ( self , *lowercase , **lowercase) -> int: '''simple docstring''' a__: int = kwargs.pop('input_values' , lowercase) a__: Optional[int] = kwargs.pop('input_ids' , lowercase) a__: int = kwargs.pop('labels' , lowercase) if input_values is not None and input_ids is not None: raise ValueError('Cannot process both `input_values` and `input_ids` inputs.') if input_values is None and input_ids is None and labels is None: raise ValueError( 'You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.') if input_values is not None: a__: str = self.feature_extractor.pad(lowercase , *lowercase , **lowercase) elif input_ids is not None: a__: Optional[int] = self.tokenizer.pad(lowercase , **lowercase) else: a__: str = None if labels is not None: if "input_ids" in labels or (isinstance(lowercase , lowercase) and "input_ids" in labels[0]): a__: Optional[int] = self.tokenizer.pad(lowercase , **lowercase) a__: int = targets['input_ids'] else: a__: Dict = self.feature_extractor.feature_size a__: str = self.feature_extractor.num_mel_bins a__: int = self.feature_extractor.pad(lowercase , *lowercase , **lowercase) a__: Optional[Any] = feature_size_hack a__: Any = targets['input_values'] else: a__: Union[str, Any] = None if inputs is None: return targets if targets is not None: a__: Any = labels a__: Optional[int] = targets.get('attention_mask') if decoder_attention_mask is not None: a__: Any = decoder_attention_mask return inputs def lowerCamelCase_ ( self , *lowercase , **lowercase) -> Optional[int]: '''simple docstring''' return self.tokenizer.batch_decode(*lowercase , **lowercase) def lowerCamelCase_ ( self , *lowercase , **lowercase) -> Optional[int]: '''simple docstring''' return self.tokenizer.decode(*lowercase , **lowercase)
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class __snake_case ( __lowerCAmelCase ): a__ = """decision_transformer""" a__ = ["""past_key_values"""] a__ = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=17 , lowercase=4 , lowercase=1_28 , lowercase=40_96 , lowercase=True , lowercase=1 , lowercase=10_24 , lowercase=3 , lowercase=1 , lowercase=None , lowercase="relu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=1e-5 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=5_02_56 , lowercase=5_02_56 , lowercase=False , lowercase=False , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = state_dim a__: int = act_dim a__: List[Any] = hidden_size a__: List[str] = max_ep_len a__: List[Any] = action_tanh a__: Optional[Any] = vocab_size a__: Tuple = n_positions a__: Dict = n_layer a__: Optional[int] = n_head a__: Optional[int] = n_inner a__: Any = activation_function a__: Union[str, Any] = resid_pdrop a__: Any = embd_pdrop a__: Any = attn_pdrop a__: List[Any] = layer_norm_epsilon a__: Optional[Any] = initializer_range a__: Any = scale_attn_weights a__: Dict = use_cache a__: Optional[int] = scale_attn_by_inverse_layer_idx a__: List[str] = reorder_and_upcast_attn a__: Any = bos_token_id a__: int = eos_token_id super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase)
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE ) ->"list[int]": if upper_limit < 0: raise ValueError('Limit for the Catalan sequence must be ≥ 0' ) a__: List[Any] = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 a__: Dict = 1 if upper_limit > 0: a__: int = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(_SCREAMING_SNAKE_CASE ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('\n********* Catalan Numbers Using Dynamic Programming ************\n') print('\n*** Enter -1 at any time to quit ***') print('\nEnter the upper limit (≥ 0) for the Catalan number sequence: ', end='') try: while True: lowercase__ = int(input().strip()) if N < 0: print('\n********* Goodbye!! ************') break else: print(f"The Catalan numbers from 0 through {N} are:") print(catalan_numbers(N)) print('Try another upper limit for the sequence: ', end='') except (NameError, ValueError): print('\n********* Invalid input, goodbye! ************\n') import doctest doctest.testmod()
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: while a != 0: a__ , a__: List[str] = b % a, a return b def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: if gcd(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) != 1: a__: Dict = F'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Union[str, Any] = 1, 0, a a__ , a__ , a__: Any = 0, 1, m while va != 0: a__: int = ua // va a__ , a__ , a__ , a__ , a__ , a__: Any = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
290
1
"""simple docstring""" import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class __snake_case : def __init__( self , lowercase = "cpu" , lowercase = "openai/clip-vit-large-patch14") -> None: '''simple docstring''' a__: List[str] = device a__: str = CLIPTokenizerFast.from_pretrained(lowercase) a__: int = [0.48145466, 0.4578275, 0.40821073] a__: Union[str, Any] = [0.26862954, 0.26130258, 0.27577711] a__: List[str] = torchvision.transforms.Normalize(self.image_mean , self.image_std) a__: Optional[int] = torchvision.transforms.Resize(2_24) a__: Optional[Any] = torchvision.transforms.CenterCrop(2_24) def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' a__: str = self.resize(lowercase) a__: str = self.center_crop(lowercase) a__: Dict = self.normalize(lowercase) return images def __call__( self , lowercase=None , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = self.tokenizer(text=lowercase , **lowercase) a__: Tuple = self.preprocess_img(lowercase) a__: List[str] = {key: value.to(self.device) for (key, value) in encoding.items()} return encoding class __snake_case ( nn.Module ): def __init__( self , lowercase=10 , lowercase=0.01 , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=False , lowercase=True , lowercase="image" , lowercase=True , lowercase=False , lowercase=False , lowercase=False , ) -> None: '''simple docstring''' super().__init__() a__: Dict = None a__: Any = device if device else get_device() if vqgan: a__: Optional[int] = vqgan else: a__: Union[str, Any] = load_vqgan(self.device , conf_path=lowercase , ckpt_path=lowercase) self.vqgan.eval() if clip: a__: Optional[int] = clip else: a__: str = CLIPModel.from_pretrained('openai/clip-vit-base-patch32') self.clip.to(self.device) a__: Tuple = ProcessorGradientFlow(device=self.device) a__: int = iterations a__: Any = lr a__: Optional[Any] = log a__: Dict = make_grid a__: int = return_val a__: Optional[int] = quantize a__: List[str] = self.vqgan.decoder.z_shape def lowerCamelCase_ ( self , lowercase=None , lowercase=None , lowercase=5 , lowercase=True) -> Any: '''simple docstring''' a__: Dict = [] if output_path is None: a__: int = './animation.gif' if input_path is None: a__: Union[str, Any] = self.save_path a__: Any = sorted(glob(input_path + '/*')) if not len(lowercase): raise ValueError( 'No images found in save path, aborting (did you pass save_intermediate=True to the generate' ' function?)') if len(lowercase) == 1: print('Only one image found in save path, (did you pass save_intermediate=True to the generate function?)') a__: Dict = total_duration / len(lowercase) a__: Optional[Any] = [frame_duration] * len(lowercase) if extend_frames: a__: Any = 1.5 a__: Any = 3 for file_name in paths: if file_name.endswith('.png'): images.append(imageio.imread(lowercase)) imageio.mimsave(lowercase , lowercase , duration=lowercase) print(f'gif saved to {output_path}') def lowerCamelCase_ ( self , lowercase=None , lowercase=None) -> Any: '''simple docstring''' if not (path or img): raise ValueError('Input either path or tensor') if img is not None: raise NotImplementedError a__: List[Any] = preprocess(Image.open(lowercase) , target_image_size=2_56).to(self.device) a__: Union[str, Any] = preprocess_vqgan(lowercase) a__ , *a__: List[Any] = self.vqgan.encode(lowercase) return z def lowerCamelCase_ ( self , lowercase) -> str: '''simple docstring''' a__: int = self.latent.detach().requires_grad_() a__: str = base_latent + transform_vector if self.quantize: a__ , *a__: int = self.vqgan.quantize(lowercase) else: a__: List[str] = trans_latent return self.vqgan.decode(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase=None) -> str: '''simple docstring''' a__: List[Any] = self.clip_preprocessor(text=lowercase , images=lowercase , return_tensors='pt' , padding=lowercase) a__: Dict = self.clip(**lowercase) a__: Any = clip_outputs.logits_per_image if weights is not None: a__: Tuple = similarity_logits * weights return similarity_logits.sum() def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> str: '''simple docstring''' a__: Dict = self._get_clip_similarity(pos_prompts['prompts'] , lowercase , weights=(1 / pos_prompts['weights'])) if neg_prompts: a__: Optional[Any] = self._get_clip_similarity(neg_prompts['prompts'] , lowercase , weights=neg_prompts['weights']) else: a__: List[Any] = torch.tensor([1] , device=self.device) a__: Dict = -torch.log(lowercase) + torch.log(lowercase) return loss def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__: Union[str, Any] = torch.randn_like(self.latent , requires_grad=lowercase , device=self.device) a__: int = torch.optim.Adam([vector] , lr=self.lr) for i in range(self.iterations): optim.zero_grad() a__: List[Any] = self._add_vector(lowercase) a__: Dict = loop_post_process(lowercase) a__: List[str] = self._get_CLIP_loss(lowercase , lowercase , lowercase) print('CLIP loss' , lowercase) if self.log: wandb.log({'CLIP Loss': clip_loss}) clip_loss.backward(retain_graph=lowercase) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0]) else: yield vector def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' wandb.init(reinit=lowercase , project='face-editor') wandb.config.update({'Positive Prompts': positive_prompts}) wandb.config.update({'Negative Prompts': negative_prompts}) wandb.config.update({'lr': self.lr, 'iterations': self.iterations}) if image_path: a__: Any = Image.open(lowercase) a__: Union[str, Any] = image.resize((2_56, 2_56)) wandb.log('Original Image' , wandb.Image(lowercase)) def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' if not prompts: return [] a__: str = [] a__: Optional[Any] = [] if isinstance(lowercase , lowercase): a__: Optional[Any] = [prompt.strip() for prompt in prompts.split('|')] for prompt in prompts: if isinstance(lowercase , (tuple, list)): a__: int = prompt[0] a__: Union[str, Any] = float(prompt[1]) elif ":" in prompt: a__ , a__: str = prompt.split(':') a__: Union[str, Any] = float(lowercase) else: a__: str = prompt a__: List[str] = 1.0 processed_prompts.append(lowercase) weights.append(lowercase) return { "prompts": processed_prompts, "weights": torch.tensor(lowercase , device=self.device), } def lowerCamelCase_ ( self , lowercase , lowercase=None , lowercase=None , lowercase=True , lowercase=False , lowercase=True , lowercase=True , lowercase=None , ) -> str: '''simple docstring''' if image_path: a__: Union[str, Any] = self._get_latent(lowercase) else: a__: List[Any] = torch.randn(self.latent_dim , device=self.device) if self.log: self._init_logging(lowercase , lowercase , lowercase) assert pos_prompts, "You must provide at least one positive prompt." a__: List[Any] = self.process_prompts(lowercase) a__: Dict = self.process_prompts(lowercase) if save_final and save_path is None: a__: Optional[int] = os.path.join('./outputs/' , '_'.join(pos_prompts['prompts'])) if not os.path.exists(lowercase): os.makedirs(lowercase) else: a__: List[Any] = save_path + '_' + get_timestamp() os.makedirs(lowercase) a__: Tuple = save_path a__: int = self.vqgan.decode(self.latent)[0] if show_intermediate: print('Original Image') show_pil(custom_to_pil(lowercase)) a__: str = loop_post_process(lowercase) for iter, transformed_img in enumerate(self._optimize_CLIP(lowercase , lowercase , lowercase)): if show_intermediate: show_pil(lowercase) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f'iter_{iter:03d}.png')) if self.log: wandb.log({'Image': wandb.Image(lowercase)}) if show_final: show_pil(lowercase) if save_final: transformed_img.save(os.path.join(self.save_path , f'iter_{iter:03d}_final.png'))
290
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex lowercase__ = logging.getLogger(__name__) class __snake_case : def __init__( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' if not self.initialized: a__: Optional[int] = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Optional[int] = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.retriever.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ , a__: str = self.retriever._main_retrieve(lowercase , lowercase) return doc_ids, retrieved_doc_embeds class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> int: '''simple docstring''' if index is not None and index.is_initialized() and len(lowercase) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ') super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Any = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase) for worker in self.retrieval_workers ]) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' logger.info('initializing retrieval') if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if len(self.retrieval_workers) > 0: # Select a random retrieval actor. a__: int = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] a__ , a__: List[Any] = ray.get(random_worker.retrieve.remote(lowercase , lowercase)) else: a__ , a__: Dict = self._main_retrieve(lowercase , lowercase) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' return super(lowercase , cls).get_tokenizers(lowercase , lowercase , **lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = kwargs.pop('config' , lowercase) or RagConfig.from_pretrained(lowercase , **lowercase) a__: Union[str, Any] = RagTokenizer.from_pretrained(lowercase , config=lowercase) a__: int = rag_tokenizer.question_encoder a__: Any = rag_tokenizer.generator if indexed_dataset is not None: a__: List[Any] = 'custom' a__: Optional[Any] = CustomHFIndex(config.retrieval_vector_size , lowercase) else: a__: Dict = cls._build_index(lowercase) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
290
1
"""simple docstring""" class __snake_case : def __init__( self , lowercase , lowercase=None , lowercase=None) -> List[str]: '''simple docstring''' a__: Dict = data a__: List[Any] = previous a__: Any = next_node def __str__( self) -> str: '''simple docstring''' return f'{self.data}' def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.data def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.next def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return self.previous class __snake_case : def __init__( self , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = head def __iter__( self) -> List[Any]: '''simple docstring''' return self def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if not self.current: raise StopIteration else: a__: Dict = self.current.get_data() a__: Optional[Any] = self.current.get_next() return value class __snake_case : def __init__( self) -> Dict: '''simple docstring''' a__: List[Any] = None # First node in list a__: Optional[int] = None # Last node in list def __str__( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.head a__: Optional[Any] = [] while current is not None: nodes.append(current.get_data()) a__: str = current.get_next() return " ".join(str(lowercase) for node in nodes) def __contains__( self , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = self.head while current: if current.get_data() == value: return True a__: Dict = current.get_next() return False def __iter__( self) -> int: '''simple docstring''' return LinkedListIterator(self.head) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: a__: Optional[Any] = node a__: Optional[Any] = node else: self.insert_before_node(self.head , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: self.set_head(lowercase) else: self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = Node(lowercase) if self.head is None: self.set_head(lowercase) else: self.set_tail(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Union[str, Any] = node a__: Optional[Any] = node.previous if node.get_previous() is None: a__: Tuple = node_to_insert else: a__: int = node_to_insert a__: Optional[int] = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Optional[int] = node a__: Tuple = node.next if node.get_next() is None: a__: Optional[int] = node_to_insert else: a__: Any = node_to_insert a__: str = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Any = 1 a__: Tuple = Node(lowercase) a__: Tuple = self.head while node: if current_position == position: self.insert_before_node(lowercase , lowercase) return current_position += 1 a__: List[Any] = node.next self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> Node: '''simple docstring''' a__: Tuple = self.head while node: if node.get_data() == item: return node a__: List[str] = node.get_next() raise Exception('Node not found') def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' if (node := self.get_node(lowercase)) is not None: if node == self.head: a__: Any = self.head.get_next() if node == self.tail: a__: List[Any] = self.tail.get_previous() self.remove_node_pointers(lowercase) @staticmethod def lowerCamelCase_ ( lowercase) -> None: '''simple docstring''' if node.get_next(): a__: Any = node.previous if node.get_previous(): a__: List[str] = node.next a__: int = None a__: Union[str, Any] = None def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self.head is None def __a ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: int = None if token is not None: a__: Tuple = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Optional[Any] = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: str = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) a__: int = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Dict = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Dict: a__: Dict = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Dict = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: List[Any] = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Optional[int] = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: List[Any] = None if token is not None: a__: Optional[int] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = result.headers['Location'] a__: Optional[int] = requests.get(_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: int = os.path.join(_SCREAMING_SNAKE_CASE , F'{artifact_name}.zip' ) with open(_SCREAMING_SNAKE_CASE , 'wb' ) as fp: fp.write(response.content ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: a__: List[Any] = [] a__: Optional[Any] = [] a__: List[Any] = None with zipfile.ZipFile(_SCREAMING_SNAKE_CASE ) as z: for filename in z.namelist(): if not os.path.isdir(_SCREAMING_SNAKE_CASE ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(_SCREAMING_SNAKE_CASE ) as f: for line in f: a__: Optional[int] = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs a__: Union[str, Any] = line[: line.index(': ' )] a__: Union[str, Any] = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed a__: Optional[int] = line[len('FAILED ' ) :] failed_tests.append(_SCREAMING_SNAKE_CASE ) elif filename == "job_name.txt": a__: Union[str, Any] = line if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( F'`errors` and `failed_tests` should have the same number of elements. Got {len(_SCREAMING_SNAKE_CASE )} for `errors` ' F'and {len(_SCREAMING_SNAKE_CASE )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some' ' problem.' ) a__: Tuple = None if job_name and job_links: a__: Dict = job_links.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # A list with elements of the form (line of error, error, failed test) a__: int = [x + [y] + [job_link] for x, y in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )] return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->str: a__: int = [] a__: Optional[int] = [os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for p in os.listdir(_SCREAMING_SNAKE_CASE ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(_SCREAMING_SNAKE_CASE , job_links=_SCREAMING_SNAKE_CASE ) ) return errors def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Any: a__: str = Counter() counter.update([x[1] for x in logs] ) a__: int = counter.most_common() a__: Any = {} for error, count in counts: if error_filter is None or error not in error_filter: a__: List[str] = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} a__: Optional[Any] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: List[str] = test.split('::' )[0] if test.startswith('tests/models/' ): a__: Dict = test.split('/' )[2] else: a__: Any = None return test def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: int = [(x[0], x[1], get_model(x[2] )) for x in logs] a__: List[Any] = [x for x in logs if x[2] is not None] a__: Optional[Any] = {x[2] for x in logs} a__: Dict = {} for test in tests: a__: Union[str, Any] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) a__: Union[str, Any] = counter.most_common() a__: List[str] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} a__: List[Any] = sum(error_counts.values() ) if n_errors > 0: a__: Any = {'count': n_errors, 'errors': error_counts} a__: Optional[int] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Any = '| no. | error | status |' a__: Any = '|-:|:-|:-|' a__: str = [header, sep] for error in reduced_by_error: a__: int = reduced_by_error[error]['count'] a__: Tuple = F'| {count} | {error[:100]} | |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->str: a__: List[str] = '| model | no. of errors | major error | count |' a__: str = '|-:|-:|-:|-:|' a__: int = [header, sep] for model in reduced_by_model: a__: Tuple = reduced_by_model[model]['count'] a__ , a__: Dict = list(reduced_by_model[model]['errors'].items() )[0] a__: Dict = F'| {model} | {count} | {error[:60]} | {_count} |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') lowercase__ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowercase__ = get_job_links(args.workflow_run_id, token=args.token) lowercase__ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowercase__ = k.find(' / ') lowercase__ = k[index + len(' / ') :] lowercase__ = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowercase__ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowercase__ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowercase__ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowercase__ = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowercase__ = reduce_by_error(errors) lowercase__ = reduce_by_model(errors) lowercase__ = make_github_table(reduced_by_error) lowercase__ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
290
1
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase__ = random.Random() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: if rng is None: a__: Any = global_rng a__: int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=4_00 , lowercase=20_00 , lowercase=1 , lowercase=0.0 , lowercase=1_60_00 , lowercase=True , lowercase=True , ) -> Union[str, Any]: '''simple docstring''' a__: Tuple = parent a__: Optional[int] = batch_size a__: Optional[Any] = min_seq_length a__: Optional[int] = max_seq_length a__: Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a__: Dict = feature_size a__: Any = padding_value a__: Optional[Any] = sampling_rate a__: Optional[Any] = return_attention_mask a__: str = do_normalize def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def lowerCamelCase_ ( self , lowercase=False , lowercase=False) -> Tuple: '''simple docstring''' def _flatten(lowercase): return list(itertools.chain(*lowercase)) if equal_length: a__: Dict = floats_list((self.batch_size, self.max_seq_length)) else: # make sure that inputs increase in size a__: List[Any] = [ _flatten(floats_list((x, self.feature_size))) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: a__: str = [np.asarray(lowercase) for x in speech_inputs] return speech_inputs class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = WavaVecaFeatureExtractor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = WavaVecaFeatureExtractionTester(self) def lowerCamelCase_ ( self , lowercase) -> List[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(lowercase , axis=0) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0) - 1) < 1e-3)) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 a__: Optional[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: List[str] = [np.asarray(lowercase) for speech_input in speech_inputs] # Test not batched input a__: Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='np').input_values a__: Dict = feat_extract(np_speech_inputs[0] , return_tensors='np').input_values self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test batched a__: Dict = feat_extract(lowercase , return_tensors='np').input_values a__: int = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test 2-D numpy arrays are batched. a__: int = [floats_list((1, x))[0] for x in (8_00, 8_00, 8_00)] a__: Union[str, Any] = np.asarray(lowercase) a__: int = feat_extract(lowercase , return_tensors='np').input_values a__: Any = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: List[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Optional[int] = ['longest', 'max_length', 'do_not_pad'] a__: List[Any] = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: Dict = feat_extract(lowercase , padding=lowercase , max_length=lowercase , return_tensors='np') a__: Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self.assertTrue(input_values[0][8_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self.assertTrue(input_values[0][10_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Optional[int] = range(8_00 , 14_00 , 2_00) a__: List[str] = [floats_list((1, x))[0] for x in lengths] a__: Tuple = ['longest', 'max_length', 'do_not_pad'] a__: Dict = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: int = feat_extract(lowercase , max_length=lowercase , padding=lowercase) a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Any = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Dict = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='max_length' , return_tensors='np') a__: int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1]) self._check_zero_mean_unit_variance(input_values[2]) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: int = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: str = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='longest' , return_tensors='np') a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00)) a__: Dict = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Tuple = feat_extract( lowercase , truncation=lowercase , max_length=20_00 , padding='longest' , return_tensors='np') a__: str = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00)) @require_torch def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' import torch a__: Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Tuple = np.random.rand(1_00).astype(np.floataa) a__: Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a__: Any = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np') self.assertTrue(np_processed.input_values.dtype == np.floataa) a__: Optional[Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt') self.assertTrue(pt_processed.input_values.dtype == torch.floataa) @slow @require_torch def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: a__: str = WavaVecaConfig.from_pretrained(lowercase) a__: str = WavaVecaFeatureExtractor.from_pretrained(lowercase) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer')
290
"""simple docstring""" import math def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __a ( _SCREAMING_SNAKE_CASE = 0.1 ) ->int: a__: str = 3 a__: Optional[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_SCREAMING_SNAKE_CASE ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class __snake_case : @staticmethod def lowerCamelCase_ ( *lowercase , **lowercase) -> int: '''simple docstring''' pass @is_pipeline_test @require_torch @require_vision class __snake_case ( unittest.TestCase ): a__ = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = pipeline('visual-question-answering' , model='hf-internal-testing/tiny-vilt-random-vqa') a__: Optional[Any] = [ { 'image': Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png'), 'question': 'How many cats are there?', }, { 'image': './tests/fixtures/tests_samples/COCO/000000039769.png', 'question': 'How many cats are there?', }, ] return vqa_pipeline, examples def lowerCamelCase_ ( self , lowercase , lowercase) -> Optional[Any]: '''simple docstring''' a__: List[str] = vqa_pipeline(lowercase , top_k=1) self.assertEqual( lowercase , [ [{'score': ANY(lowercase), 'answer': ANY(lowercase)}], [{'score': ANY(lowercase), 'answer': ANY(lowercase)}], ] , ) @require_torch def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: List[Any] = pipeline('visual-question-answering' , model='hf-internal-testing/tiny-vilt-random-vqa') a__: Optional[int] = './tests/fixtures/tests_samples/COCO/000000039769.png' a__: Optional[Any] = 'How many cats are there?' a__: List[Any] = vqa_pipeline(image=lowercase , question='How many cats are there?' , top_k=2) self.assertEqual( lowercase , [{'score': ANY(lowercase), 'answer': ANY(lowercase)}, {'score': ANY(lowercase), 'answer': ANY(lowercase)}]) a__: Tuple = vqa_pipeline({'image': image, 'question': question} , top_k=2) self.assertEqual( lowercase , [{'score': ANY(lowercase), 'answer': ANY(lowercase)}, {'score': ANY(lowercase), 'answer': ANY(lowercase)}]) @slow @require_torch def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = pipeline('visual-question-answering' , model='dandelin/vilt-b32-finetuned-vqa') a__: Optional[int] = './tests/fixtures/tests_samples/COCO/000000039769.png' a__: Dict = 'How many cats are there?' a__: List[Any] = vqa_pipeline(image=lowercase , question=lowercase , top_k=2) self.assertEqual( nested_simplify(lowercase , decimals=4) , [{'score': 0.8799, 'answer': '2'}, {'score': 0.296, 'answer': '1'}]) a__: Tuple = vqa_pipeline({'image': image, 'question': question} , top_k=2) self.assertEqual( nested_simplify(lowercase , decimals=4) , [{'score': 0.8799, 'answer': '2'}, {'score': 0.296, 'answer': '1'}]) a__: Tuple = vqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2) self.assertEqual( nested_simplify(lowercase , decimals=4) , [[{'score': 0.8799, 'answer': '2'}, {'score': 0.296, 'answer': '1'}]] * 2 , ) @require_tf @unittest.skip('Visual question answering not implemented in TF') def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' pass
290
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ = {'processing_wav2vec2_with_lm': ['Wav2Vec2ProcessorWithLM']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
1
"""simple docstring""" import os import unittest from transformers import MobileBertTokenizer, MobileBertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = MobileBertTokenizer a__ = MobileBertTokenizerFast a__ = True a__ = True a__ = filter_non_english a__ = """google/mobilebert-uncased""" def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().setUp() a__: Optional[int] = [ '[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] a__: Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens])) a__: Dict = [ (tokenizer_def[0], self.pre_trained_model_path, tokenizer_def[2]) # else the 'google/' prefix is stripped for tokenizer_def in self.tokenizers_list ] def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' a__: List[Any] = 'UNwant\u00E9d,running' a__: Dict = 'unwanted, running' return input_text, output_text def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: int = self.tokenizer_class(self.vocab_file) a__: Tuple = tokenizer.tokenize('UNwant\u00E9d,running') self.assertListEqual(lowercase , ['un', '##want', '##ed', ',', 'runn', '##ing']) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase) , [9, 6, 7, 12, 10, 11]) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if not self.test_rust_tokenizer: return a__: List[Any] = self.get_tokenizer() a__: Dict = self.get_rust_tokenizer() a__: Dict = 'UNwant\u00E9d,running' a__: List[str] = tokenizer.tokenize(lowercase) a__: Optional[int] = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) a__: int = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: List[str] = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) a__: Dict = self.get_rust_tokenizer() a__: str = tokenizer.encode(lowercase) a__: Dict = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) # With lower casing a__: str = self.get_tokenizer(do_lower_case=lowercase) a__: Tuple = self.get_rust_tokenizer(do_lower_case=lowercase) a__: List[Any] = 'UNwant\u00E9d,running' a__: Optional[Any] = tokenizer.tokenize(lowercase) a__: int = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) a__: List[Any] = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) a__: Tuple = self.get_rust_tokenizer() a__: Any = tokenizer.encode(lowercase) a__: int = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = BasicTokenizer() self.assertListEqual(tokenizer.tokenize('ah\u535A\u63A8zz') , ['ah', '\u535A', '\u63A8', 'zz']) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Optional[Any] = BasicTokenizer(do_lower_case=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ') , ['hello', '!', 'how', 'are', 'you', '?']) self.assertListEqual(tokenizer.tokenize('H\u00E9llo') , ['hello']) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[Any] = BasicTokenizer(do_lower_case=lowercase , strip_accents=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ') , ['hällo', '!', 'how', 'are', 'you', '?']) self.assertListEqual(tokenizer.tokenize('H\u00E9llo') , ['h\u00E9llo']) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: Dict = BasicTokenizer(do_lower_case=lowercase , strip_accents=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ') , ['hallo', '!', 'how', 'are', 'you', '?']) self.assertListEqual(tokenizer.tokenize('H\u00E9llo') , ['hello']) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[str] = BasicTokenizer(do_lower_case=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ') , ['hallo', '!', 'how', 'are', 'you', '?']) self.assertListEqual(tokenizer.tokenize('H\u00E9llo') , ['hello']) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: List[Any] = BasicTokenizer(do_lower_case=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? ') , ['HeLLo', '!', 'how', 'Are', 'yoU', '?']) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: str = BasicTokenizer(do_lower_case=lowercase , strip_accents=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ') , ['HäLLo', '!', 'how', 'Are', 'yoU', '?']) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: str = BasicTokenizer(do_lower_case=lowercase , strip_accents=lowercase) self.assertListEqual( tokenizer.tokenize(' \tHäLLo!how \n Are yoU? ') , ['HaLLo', '!', 'how', 'Are', 'yoU', '?']) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Dict = BasicTokenizer(do_lower_case=lowercase , never_split=['[UNK]']) self.assertListEqual( tokenizer.tokenize(' \tHeLLo!how \n Are yoU? [UNK]') , ['HeLLo', '!', 'how', 'Are', 'yoU', '?', '[UNK]']) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Tuple = ['[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing'] a__: List[str] = {} for i, token in enumerate(lowercase): a__: List[str] = i a__: Optional[Any] = WordpieceTokenizer(vocab=lowercase , unk_token='[UNK]') self.assertListEqual(tokenizer.tokenize('') , []) self.assertListEqual(tokenizer.tokenize('unwanted running') , ['un', '##want', '##ed', 'runn', '##ing']) self.assertListEqual(tokenizer.tokenize('unwantedX running') , ['[UNK]', 'runn', '##ing']) def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.assertTrue(_is_whitespace(' ')) self.assertTrue(_is_whitespace('\t')) self.assertTrue(_is_whitespace('\r')) self.assertTrue(_is_whitespace('\n')) self.assertTrue(_is_whitespace('\u00A0')) self.assertFalse(_is_whitespace('A')) self.assertFalse(_is_whitespace('-')) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' self.assertTrue(_is_control('\u0005')) self.assertFalse(_is_control('A')) self.assertFalse(_is_control(' ')) self.assertFalse(_is_control('\t')) self.assertFalse(_is_control('\r')) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' self.assertTrue(_is_punctuation('-')) self.assertTrue(_is_punctuation('$')) self.assertTrue(_is_punctuation('`')) self.assertTrue(_is_punctuation('.')) self.assertFalse(_is_punctuation('A')) self.assertFalse(_is_punctuation(' ')) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = self.get_tokenizer() a__: Any = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(lowercase) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']]) self.assertListEqual( [rust_tokenizer.tokenize(lowercase) for t in ['Test', '\xad', 'test']] , [['[UNK]'], [], ['[UNK]']]) @slow def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[int] = self.tokenizer_class.from_pretrained('google/mobilebert-uncased') a__: List[Any] = tokenizer.encode('sequence builders' , add_special_tokens=lowercase) a__: Tuple = tokenizer.encode('multi-sequence build' , add_special_tokens=lowercase) a__: Any = tokenizer.build_inputs_with_special_tokens(lowercase) a__: int = tokenizer.build_inputs_with_special_tokens(lowercase , lowercase) assert encoded_sentence == [1_01] + text + [1_02] assert encoded_pair == [1_01] + text + [1_02] + text_a + [1_02] def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})'): a__: Optional[Any] = self.rust_tokenizer_class.from_pretrained(lowercase , **lowercase) a__: str = f'A, naïve {tokenizer_r.mask_token} AllenNLP sentence.' a__: Optional[Any] = tokenizer_r.encode_plus( lowercase , return_attention_mask=lowercase , return_token_type_ids=lowercase , return_offsets_mapping=lowercase , add_special_tokens=lowercase , ) a__: int = tokenizer_r.do_lower_case if hasattr(lowercase , 'do_lower_case') else False a__: int = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'A'), ((1, 2), ','), ((3, 5), 'na'), ((5, 6), '##ï'), ((6, 8), '##ve'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'Allen'), ((21, 23), '##NL'), ((23, 24), '##P'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), 'a'), ((1, 2), ','), ((3, 8), 'naive'), ((9, 15), tokenizer_r.mask_token), ((16, 21), 'allen'), ((21, 23), '##nl'), ((23, 24), '##p'), ((25, 33), 'sentence'), ((33, 34), '.'), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens['input_ids'])) self.assertEqual([e[0] for e in expected_results] , tokens['offset_mapping']) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Any = ['的', '人', '有'] a__: Union[str, Any] = ''.join(lowercase) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})'): a__: Optional[int] = True a__: Union[str, Any] = self.tokenizer_class.from_pretrained(lowercase , **lowercase) a__: List[Any] = self.rust_tokenizer_class.from_pretrained(lowercase , **lowercase) a__: int = tokenizer_p.encode(lowercase , add_special_tokens=lowercase) a__: Union[str, Any] = tokenizer_r.encode(lowercase , add_special_tokens=lowercase) a__: List[Any] = tokenizer_r.convert_ids_to_tokens(lowercase) a__: List[Any] = tokenizer_p.convert_ids_to_tokens(lowercase) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(lowercase , lowercase) self.assertListEqual(lowercase , lowercase) a__: Any = False a__: Dict = self.rust_tokenizer_class.from_pretrained(lowercase , **lowercase) a__: List[Any] = self.tokenizer_class.from_pretrained(lowercase , **lowercase) a__: str = tokenizer_r.encode(lowercase , add_special_tokens=lowercase) a__: str = tokenizer_p.encode(lowercase , add_special_tokens=lowercase) a__: Tuple = tokenizer_r.convert_ids_to_tokens(lowercase) a__: List[Any] = tokenizer_p.convert_ids_to_tokens(lowercase) # it is expected that only the first Chinese character is not preceded by "##". a__: Dict = [ f'##{token}' if idx != 0 else token for idx, token in enumerate(lowercase) ] self.assertListEqual(lowercase , lowercase) self.assertListEqual(lowercase , lowercase)
290
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = KandinskyInpaintPipeline a__ = ["""prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] a__ = [ """prompt""", """negative_prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] a__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """negative_prompt""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] a__ = False @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 1_00 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[int] = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base') return tokenizer @property def lowerCamelCase_ ( self) -> Any: '''simple docstring''' torch.manual_seed(0) a__: Dict = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=10_05 , ) a__: Optional[Any] = MultilingualCLIP(lowercase) a__: int = text_encoder.eval() return text_encoder @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' torch.manual_seed(0) a__: Any = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } a__: str = UNetaDConditionModel(**lowercase) return model @property def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' torch.manual_seed(0) a__: Any = VQModel(**self.dummy_movq_kwargs) return model def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = self.dummy_text_encoder a__: int = self.dummy_tokenizer a__: str = self.dummy_unet a__: Any = self.dummy_movq a__: Tuple = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.00085 , beta_end=0.012 , clip_sample=lowercase , set_alpha_to_one=lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=lowercase , ) a__: Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> Any: '''simple docstring''' a__: List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase)).to(lowercase) a__: int = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1)).to(lowercase) # create init_image a__: Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) a__: int = image.cpu().permute(0 , 2 , 3 , 1)[0] a__: Optional[int] = Image.fromarray(np.uinta(lowercase)).convert('RGB').resize((2_56, 2_56)) # create mask a__: Tuple = np.ones((64, 64) , dtype=np.floataa) a__: Optional[Any] = 0 if str(lowercase).startswith('mps'): a__: str = torch.manual_seed(lowercase) else: a__: Dict = torch.Generator(device=lowercase).manual_seed(lowercase) a__: Optional[int] = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[Any] = 'cpu' a__: List[Any] = self.get_dummy_components() a__: Optional[Any] = self.pipeline_class(**lowercase) a__: str = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[int] = pipe(**self.get_dummy_inputs(lowercase)) a__: List[str] = output.images a__: int = pipe( **self.get_dummy_inputs(lowercase) , return_dict=lowercase , )[0] a__: Optional[Any] = image[0, -3:, -3:, -1] a__: List[Any] = image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}') assert image.shape == (1, 64, 64, 3) a__: str = np.array( [0.8326919, 0.73790467, 0.20918581, 0.9309612, 0.5511791, 0.43713328, 0.5513321, 0.49922934, 0.59497786]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy') a__: int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png') a__: Union[str, Any] = np.ones((7_68, 7_68) , dtype=np.floataa) a__: int = 0 a__: Optional[int] = 'a hat' a__: int = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa) pipe_prior.to(lowercase) a__: Any = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa) a__: Optional[Any] = pipeline.to(lowercase) pipeline.set_progress_bar_config(disable=lowercase) a__: Dict = torch.Generator(device='cpu').manual_seed(0) a__ , a__: Optional[Any] = pipe_prior( lowercase , generator=lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() a__: List[str] = pipeline( lowercase , image=lowercase , mask_image=lowercase , image_embeds=lowercase , negative_image_embeds=lowercase , generator=lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) a__: str = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(lowercase , lowercase)
290
1
"""simple docstring""" import math def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __a ( _SCREAMING_SNAKE_CASE = 0.1 ) ->int: a__: str = 3 a__: Optional[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_SCREAMING_SNAKE_CASE ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() lowercase__ = logging.get_logger('transformers.models.encodec') lowercase__ = { 'quantizer.vq.layers.*._codebook.inited': 'quantizer.layers.*.codebook.inited', 'quantizer.vq.layers.*._codebook.cluster_size': 'quantizer.layers.*.codebook.cluster_size', 'quantizer.vq.layers.*._codebook.embed': 'quantizer.layers.*.codebook.embed', 'quantizer.vq.layers.*._codebook.embed_avg': 'quantizer.layers.*.codebook.embed_avg', } lowercase__ = { 'encoder.model.0.conv.conv': 'encoder.layers.0.conv', 'encoder.model.1.block.1.conv.conv': 'encoder.layers.1.block.1.conv', 'encoder.model.1.block.3.conv.conv': 'encoder.layers.1.block.3.conv', 'encoder.model.1.shortcut.conv.conv': 'encoder.layers.1.shortcut.conv', 'encoder.model.3.conv.conv': 'encoder.layers.3.conv', 'encoder.model.4.block.1.conv.conv': 'encoder.layers.4.block.1.conv', 'encoder.model.4.block.3.conv.conv': 'encoder.layers.4.block.3.conv', 'encoder.model.4.shortcut.conv.conv': 'encoder.layers.4.shortcut.conv', 'encoder.model.6.conv.conv': 'encoder.layers.6.conv', 'encoder.model.7.block.1.conv.conv': 'encoder.layers.7.block.1.conv', 'encoder.model.7.block.3.conv.conv': 'encoder.layers.7.block.3.conv', 'encoder.model.7.shortcut.conv.conv': 'encoder.layers.7.shortcut.conv', 'encoder.model.9.conv.conv': 'encoder.layers.9.conv', 'encoder.model.10.block.1.conv.conv': 'encoder.layers.10.block.1.conv', 'encoder.model.10.block.3.conv.conv': 'encoder.layers.10.block.3.conv', 'encoder.model.10.shortcut.conv.conv': 'encoder.layers.10.shortcut.conv', 'encoder.model.12.conv.conv': 'encoder.layers.12.conv', 'encoder.model.13.lstm': 'encoder.layers.13.lstm', 'encoder.model.15.conv.conv': 'encoder.layers.15.conv', } lowercase__ = { 'encoder.model.0.conv.norm': 'encoder.layers.0.norm', 'encoder.model.1.block.1.conv.norm': 'encoder.layers.1.block.1.norm', 'encoder.model.1.block.3.conv.norm': 'encoder.layers.1.block.3.norm', 'encoder.model.1.shortcut.conv.norm': 'encoder.layers.1.shortcut.norm', 'encoder.model.3.conv.norm': 'encoder.layers.3.norm', 'encoder.model.4.block.1.conv.norm': 'encoder.layers.4.block.1.norm', 'encoder.model.4.block.3.conv.norm': 'encoder.layers.4.block.3.norm', 'encoder.model.4.shortcut.conv.norm': 'encoder.layers.4.shortcut.norm', 'encoder.model.6.conv.norm': 'encoder.layers.6.norm', 'encoder.model.7.block.1.conv.norm': 'encoder.layers.7.block.1.norm', 'encoder.model.7.block.3.conv.norm': 'encoder.layers.7.block.3.norm', 'encoder.model.7.shortcut.conv.norm': 'encoder.layers.7.shortcut.norm', 'encoder.model.9.conv.norm': 'encoder.layers.9.norm', 'encoder.model.10.block.1.conv.norm': 'encoder.layers.10.block.1.norm', 'encoder.model.10.block.3.conv.norm': 'encoder.layers.10.block.3.norm', 'encoder.model.10.shortcut.conv.norm': 'encoder.layers.10.shortcut.norm', 'encoder.model.12.conv.norm': 'encoder.layers.12.norm', 'encoder.model.15.conv.norm': 'encoder.layers.15.norm', } lowercase__ = { 'decoder.model.0.conv.conv': 'decoder.layers.0.conv', 'decoder.model.1.lstm': 'decoder.layers.1.lstm', 'decoder.model.3.convtr.convtr': 'decoder.layers.3.conv', 'decoder.model.4.block.1.conv.conv': 'decoder.layers.4.block.1.conv', 'decoder.model.4.block.3.conv.conv': 'decoder.layers.4.block.3.conv', 'decoder.model.4.shortcut.conv.conv': 'decoder.layers.4.shortcut.conv', 'decoder.model.6.convtr.convtr': 'decoder.layers.6.conv', 'decoder.model.7.block.1.conv.conv': 'decoder.layers.7.block.1.conv', 'decoder.model.7.block.3.conv.conv': 'decoder.layers.7.block.3.conv', 'decoder.model.7.shortcut.conv.conv': 'decoder.layers.7.shortcut.conv', 'decoder.model.9.convtr.convtr': 'decoder.layers.9.conv', 'decoder.model.10.block.1.conv.conv': 'decoder.layers.10.block.1.conv', 'decoder.model.10.block.3.conv.conv': 'decoder.layers.10.block.3.conv', 'decoder.model.10.shortcut.conv.conv': 'decoder.layers.10.shortcut.conv', 'decoder.model.12.convtr.convtr': 'decoder.layers.12.conv', 'decoder.model.13.block.1.conv.conv': 'decoder.layers.13.block.1.conv', 'decoder.model.13.block.3.conv.conv': 'decoder.layers.13.block.3.conv', 'decoder.model.13.shortcut.conv.conv': 'decoder.layers.13.shortcut.conv', 'decoder.model.15.conv.conv': 'decoder.layers.15.conv', } lowercase__ = { 'decoder.model.0.conv.norm': 'decoder.layers.0.norm', 'decoder.model.3.convtr.norm': 'decoder.layers.3.norm', 'decoder.model.4.block.1.conv.norm': 'decoder.layers.4.block.1.norm', 'decoder.model.4.block.3.conv.norm': 'decoder.layers.4.block.3.norm', 'decoder.model.4.shortcut.conv.norm': 'decoder.layers.4.shortcut.norm', 'decoder.model.6.convtr.norm': 'decoder.layers.6.norm', 'decoder.model.7.block.1.conv.norm': 'decoder.layers.7.block.1.norm', 'decoder.model.7.block.3.conv.norm': 'decoder.layers.7.block.3.norm', 'decoder.model.7.shortcut.conv.norm': 'decoder.layers.7.shortcut.norm', 'decoder.model.9.convtr.norm': 'decoder.layers.9.norm', 'decoder.model.10.block.1.conv.norm': 'decoder.layers.10.block.1.norm', 'decoder.model.10.block.3.conv.norm': 'decoder.layers.10.block.3.norm', 'decoder.model.10.shortcut.conv.norm': 'decoder.layers.10.shortcut.norm', 'decoder.model.12.convtr.norm': 'decoder.layers.12.norm', 'decoder.model.13.block.1.conv.norm': 'decoder.layers.13.block.1.norm', 'decoder.model.13.block.3.conv.norm': 'decoder.layers.13.block.3.norm', 'decoder.model.13.shortcut.conv.norm': 'decoder.layers.13.shortcut.norm', 'decoder.model.15.conv.norm': 'decoder.layers.15.norm', } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } lowercase__ = [] lowercase__ = [] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: for attribute in key.split('.' ): a__: str = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: str = value elif weight_type == "weight_g": a__: int = value elif weight_type == "weight_v": a__: Tuple = value elif weight_type == "bias": a__: Dict = value elif weight_type == "running_mean": a__: Any = value elif weight_type == "running_var": a__: Tuple = value elif weight_type == "num_batches_tracked": a__: List[str] = value elif weight_type == "weight_ih_l0": a__: List[Any] = value elif weight_type == "weight_hh_l0": a__: List[Any] = value elif weight_type == "bias_ih_l0": a__: List[Any] = value elif weight_type == "bias_hh_l0": a__: List[Any] = value elif weight_type == "weight_ih_l1": a__: int = value elif weight_type == "weight_hh_l1": a__: str = value elif weight_type == "bias_ih_l1": a__: Union[str, Any] = value elif weight_type == "bias_hh_l1": a__: Any = value else: a__: Union[str, Any] = value logger.info(F'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: for key in ignore_keys: if key.endswith('.*' ): if name.startswith(key[:-1] ): return True elif ".*." in key: a__ , a__: Optional[Any] = key.split('.*.' ) if prefix in name and suffix in name: return True elif key in name: return True return False def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[Any] = [] if model_name == "encodec_24khz" or "encodec_32khz": a__: Optional[int] = MAPPING_24K elif model_name == "encodec_48khz": a__: List[Any] = MAPPING_48K else: raise ValueError(F'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): logger.info(F'{name} was ignored' ) continue a__: int = False for key, mapped_key in MAPPING.items(): if "*" in key: a__ , a__: str = key.split('.*.' ) if prefix in name and suffix in name: a__: List[str] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('embed' ) and name.endswith('embed_avg' ): continue a__: List[str] = True if "*" in mapped_key: a__: List[str] = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: str = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: int = 'weight_g' elif "weight_v" in name: a__: Dict = 'weight_v' elif "weight_ih_l0" in name: a__: int = 'weight_ih_l0' elif "weight_hh_l0" in name: a__: Union[str, Any] = 'weight_hh_l0' elif "bias_ih_l0" in name: a__: Optional[Any] = 'bias_ih_l0' elif "bias_hh_l0" in name: a__: Optional[int] = 'bias_hh_l0' elif "weight_ih_l1" in name: a__: Dict = 'weight_ih_l1' elif "weight_hh_l1" in name: a__: Optional[Any] = 'weight_hh_l1' elif "bias_ih_l1" in name: a__: List[str] = 'bias_ih_l1' elif "bias_hh_l1" in name: a__: Optional[Any] = 'bias_hh_l1' elif "bias" in name: a__: List[str] = 'bias' elif "weight" in name: a__: Any = 'weight' elif "running_mean" in name: a__: Dict = 'running_mean' elif "running_var" in name: a__: Dict = 'running_var' elif "num_batches_tracked" in name: a__: Dict = 'num_batches_tracked' else: a__: List[str] = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->int: if config_path is not None: a__: Dict = EncodecConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: a__: Tuple = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": a__: Any = [8, 5, 4, 4] a__: List[str] = [2.2] a__: List[Any] = 64 a__: Dict = 32000 a__: Union[str, Any] = 2048 a__: Union[str, Any] = False a__: Any = False a__: Optional[Any] = False elif model_name == "encodec_48khz": a__: Optional[int] = [8, 5, 4, 2] a__: Union[str, Any] = [3.0, 6.0, 12.0, 24.0] a__: List[str] = 48000 a__: Tuple = 2 a__: Optional[Any] = False a__: Optional[int] = 'time_group_norm' a__: Union[str, Any] = True a__: Dict = 1.0 a__: str = 0.01 else: raise ValueError(F'Unknown model name: {model_name}' ) a__: Optional[int] = EncodecModel(_SCREAMING_SNAKE_CASE ) a__: List[str] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: int = torch.load(_SCREAMING_SNAKE_CASE ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights a__: str = original_checkpoint['best_state'] recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if repo_id: print('Pushing to the hub...' ) feature_extractor.push_to_hub(_SCREAMING_SNAKE_CASE ) model.push_to_hub(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE ) ->bool: a__: set[int] = set() # To detect a back edge, keep track of vertices currently in the recursion stack a__: set[int] = set() return any( node not in visited and depth_first_search(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for node in graph ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->bool: visited.add(_SCREAMING_SNAKE_CASE ) rec_stk.add(_SCREAMING_SNAKE_CASE ) for node in graph[vertex]: if node not in visited: if depth_first_search(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(_SCREAMING_SNAKE_CASE ) return False if __name__ == "__main__": from doctest import testmod testmod()
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: print('moving disk from' , _SCREAMING_SNAKE_CASE , 'to' , _SCREAMING_SNAKE_CASE ) def __a ( ) ->List[str]: a__: Dict = int(input('Height of hanoi: ' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE = 4000000 ) ->int: a__: Optional[Any] = [] a__ , a__: List[str] = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(_SCREAMING_SNAKE_CASE ) a__ , a__: Union[str, Any] = b, a + b return sum(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(f"{solution() = }")
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = False ) ->str: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = F'Expected string as input, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[str] = F'Expected boolean as use_pascal parameter, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) a__: int = input_str.split('_' ) a__: List[str] = 0 if use_pascal else 1 a__: List[str] = words[start_index:] a__: List[str] = [word[0].upper() + word[1:] for word in words_to_capitalize] a__: List[str] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) lowercase__ = { 'configuration_mega': ['MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MegaConfig', 'MegaOnnxConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'MEGA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MegaForCausalLM', 'MegaForMaskedLM', 'MegaForMultipleChoice', 'MegaForQuestionAnswering', 'MegaForSequenceClassification', 'MegaForTokenClassification', 'MegaModel', 'MegaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mega import MEGA_PRETRAINED_CONFIG_ARCHIVE_MAP, MegaConfig, MegaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mega import ( MEGA_PRETRAINED_MODEL_ARCHIVE_LIST, MegaForCausalLM, MegaForMaskedLM, MegaForMultipleChoice, MegaForQuestionAnswering, MegaForSequenceClassification, MegaForTokenClassification, MegaModel, MegaPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" class __snake_case : def __init__( self , lowercase , lowercase=None , lowercase=None) -> List[str]: '''simple docstring''' a__: Dict = data a__: List[Any] = previous a__: Any = next_node def __str__( self) -> str: '''simple docstring''' return f'{self.data}' def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.data def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.next def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return self.previous class __snake_case : def __init__( self , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = head def __iter__( self) -> List[Any]: '''simple docstring''' return self def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if not self.current: raise StopIteration else: a__: Dict = self.current.get_data() a__: Optional[Any] = self.current.get_next() return value class __snake_case : def __init__( self) -> Dict: '''simple docstring''' a__: List[Any] = None # First node in list a__: Optional[int] = None # Last node in list def __str__( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.head a__: Optional[Any] = [] while current is not None: nodes.append(current.get_data()) a__: str = current.get_next() return " ".join(str(lowercase) for node in nodes) def __contains__( self , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = self.head while current: if current.get_data() == value: return True a__: Dict = current.get_next() return False def __iter__( self) -> int: '''simple docstring''' return LinkedListIterator(self.head) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: a__: Optional[Any] = node a__: Optional[Any] = node else: self.insert_before_node(self.head , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: self.set_head(lowercase) else: self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = Node(lowercase) if self.head is None: self.set_head(lowercase) else: self.set_tail(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Union[str, Any] = node a__: Optional[Any] = node.previous if node.get_previous() is None: a__: Tuple = node_to_insert else: a__: int = node_to_insert a__: Optional[int] = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Optional[int] = node a__: Tuple = node.next if node.get_next() is None: a__: Optional[int] = node_to_insert else: a__: Any = node_to_insert a__: str = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Any = 1 a__: Tuple = Node(lowercase) a__: Tuple = self.head while node: if current_position == position: self.insert_before_node(lowercase , lowercase) return current_position += 1 a__: List[Any] = node.next self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> Node: '''simple docstring''' a__: Tuple = self.head while node: if node.get_data() == item: return node a__: List[str] = node.get_next() raise Exception('Node not found') def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' if (node := self.get_node(lowercase)) is not None: if node == self.head: a__: Any = self.head.get_next() if node == self.tail: a__: List[Any] = self.tail.get_previous() self.remove_node_pointers(lowercase) @staticmethod def lowerCamelCase_ ( lowercase) -> None: '''simple docstring''' if node.get_next(): a__: Any = node.previous if node.get_previous(): a__: List[str] = node.next a__: int = None a__: Union[str, Any] = None def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self.head is None def __a ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" from math import sqrt def __a ( _SCREAMING_SNAKE_CASE ) ->int: a__: List[Any] = 0 for i in range(1 , int(sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) ): if n % i == 0 and i != sqrt(_SCREAMING_SNAKE_CASE ): total += i + n // i elif i == sqrt(_SCREAMING_SNAKE_CASE ): total += i return total - n def __a ( _SCREAMING_SNAKE_CASE = 10000 ) ->int: a__: Optional[Any] = sum( i for i in range(1 , _SCREAMING_SNAKE_CASE ) if sum_of_divisors(sum_of_divisors(_SCREAMING_SNAKE_CASE ) ) == i and sum_of_divisors(_SCREAMING_SNAKE_CASE ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
290
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
1
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 16 , _SCREAMING_SNAKE_CASE = "bert-base-cased" ) ->Tuple: a__: Optional[int] = AutoTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ) a__: str = load_dataset('glue' , 'mrpc' ) def tokenize_function(_SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) a__: Optional[Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset a__: Optional[Any] = datasets.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=_SCREAMING_SNAKE_CASE ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library a__: Optional[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(_SCREAMING_SNAKE_CASE , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(_SCREAMING_SNAKE_CASE , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. a__: Any = DataLoader( tokenized_datasets['train'] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) a__: Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=_SCREAMING_SNAKE_CASE , collate_fn=_SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: # Initialize accelerator a__: List[str] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs a__: int = config['lr'] a__: str = int(config['num_epochs'] ) a__: str = int(config['seed'] ) a__: Tuple = int(config['batch_size'] ) a__: List[str] = args.model_name_or_path set_seed(_SCREAMING_SNAKE_CASE ) a__ , a__: List[str] = get_dataloaders(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) a__: str = AutoModelForSequenceClassification.from_pretrained(_SCREAMING_SNAKE_CASE , return_dict=_SCREAMING_SNAKE_CASE ) # Instantiate optimizer a__: Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) a__: Optional[int] = optimizer_cls(params=model.parameters() , lr=_SCREAMING_SNAKE_CASE ) if accelerator.state.deepspeed_plugin is not None: a__: Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: a__: str = 1 a__: Union[str, Any] = (len(_SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): a__: Tuple = get_linear_schedule_with_warmup( optimizer=_SCREAMING_SNAKE_CASE , num_warmup_steps=0 , num_training_steps=_SCREAMING_SNAKE_CASE , ) else: a__: Tuple = DummyScheduler(_SCREAMING_SNAKE_CASE , total_num_steps=_SCREAMING_SNAKE_CASE , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. a__ , a__ , a__ , a__ , a__: Union[str, Any] = accelerator.prepare( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # We need to keep track of how many total steps we have iterated over a__: List[str] = 0 # We also need to keep track of the stating epoch so files are named properly a__: Optional[int] = 0 # Now we train the model a__: str = evaluate.load('glue' , 'mrpc' ) a__: Optional[Any] = 0 a__: str = {} for epoch in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): a__: List[Any] = model(**_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = outputs.loss a__: str = loss / gradient_accumulation_steps accelerator.backward(_SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() a__: List[str] = 0 for step, batch in enumerate(_SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): a__: int = model(**_SCREAMING_SNAKE_CASE ) a__: Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times a__ , a__: List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(_SCREAMING_SNAKE_CASE ) - 1: a__: Any = predictions[: len(eval_dataloader.dataset ) - samples_seen] a__: Union[str, Any] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=_SCREAMING_SNAKE_CASE , references=_SCREAMING_SNAKE_CASE , ) a__: Union[str, Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}:' , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: a__: Optional[Any] = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), F'Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( ) ->int: a__: str = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=_SCREAMING_SNAKE_CASE , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=_SCREAMING_SNAKE_CASE , ) parser.add_argument( '--output_dir' , type=_SCREAMING_SNAKE_CASE , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=_SCREAMING_SNAKE_CASE , default=3 , help='Number of train epochs.' , ) a__: List[Any] = parser.parse_args() a__: Any = {'lr': 2e-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
290
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowercase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Any: if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): a__: Optional[int] = [image] a__: str = [trans(img.convert('RGB' ) ) for img in image] a__: Any = torch.stack(_SCREAMING_SNAKE_CASE ) return image class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM a__: Dict = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowercase , scheduler=lowercase) def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f'The value of strength should in [0.0, 1.0] but is {strength}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: int = min(int(num_inference_steps * strength) , lowercase) a__: Any = max(num_inference_steps - init_timestep , 0) a__: Union[str, Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> List[Any]: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( f'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase)}') a__: Tuple = image.to(device=lowercase , dtype=lowercase) if isinstance(lowercase , lowercase) and len(lowercase) != batch_size: raise ValueError( f'You have passed a list of generators of length {len(lowercase)}, but requested an effective batch' f' size of {batch_size}. Make sure the batch size matches the length of the generators.') a__: List[str] = init_latents.shape a__: List[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) # get latents print('add noise to latents at timestep' , lowercase) a__: int = self.scheduler.add_noise(lowercase , lowercase , lowercase) a__: Dict = init_latents return latents @torch.no_grad() def __call__( self , lowercase = None , lowercase = 0.8 , lowercase = 1 , lowercase = None , lowercase = 0.0 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase) # 2. Preprocess image a__: Tuple = preprocess(lowercase) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device) a__ , a__: Union[str, Any] = self.get_timesteps(lowercase , lowercase , self.device) a__: Optional[int] = timesteps[:1].repeat(lowercase) # 4. Prepare latent variables a__: Union[str, Any] = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase) a__: Optional[Any] = latents # 5. Denoising loop for t in self.progress_bar(lowercase): # 1. predict noise model_output a__: Dict = self.unet(lowercase , lowercase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample a__: Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1) a__: Optional[int] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": a__: Dict = self.numpy_to_pil(lowercase) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase)
290
1
"""simple docstring""" from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput lowercase__ = 8 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ) ->Optional[Any]: a__: Any = x.device a__: List[str] = (x * 255).int().clamp(0 , 255 ) a__: Union[str, Any] = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE ) a__: List[str] = rearrange(_SCREAMING_SNAKE_CASE , 'd -> d 1 1' ) a__: int = rearrange(_SCREAMING_SNAKE_CASE , 'b c h w -> b c 1 h w' ) a__: Dict = ((x & mask) != 0).float() a__: Optional[Any] = rearrange(_SCREAMING_SNAKE_CASE , 'b c d h w -> b (c d) h w' ) a__: Optional[Any] = bits * 2 - 1 return bits def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=BITS ) ->Dict: a__: Optional[int] = x.device a__: List[Any] = (x > 0).int() a__: Any = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_SCREAMING_SNAKE_CASE , dtype=torch.intaa ) a__: Optional[int] = rearrange(_SCREAMING_SNAKE_CASE , 'd -> d 1 1' ) a__: str = rearrange(_SCREAMING_SNAKE_CASE , 'b (c d) h w -> b c d h w' , d=8 ) a__: Optional[int] = reduce(x * mask , 'b c d h w -> b c h w' , 'sum' ) return (dec / 255).clamp(0.0 , 1.0 ) def __a ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ) ->Union[DDIMSchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) a__: Tuple = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas a__: List[Any] = self.alphas_cumprod[timestep] a__: int = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod a__: Optional[Any] = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf a__: List[str] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" a__: Any = self.bit_scale if self.config.clip_sample: a__: List[Any] = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) a__: int = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Optional[int] = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide a__: Dict = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf a__: str = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf a__: Union[str, Any] = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 a__: Tuple = model_output.device if torch.is_tensor(_SCREAMING_SNAKE_CASE ) else 'cpu' a__: List[str] = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) a__: List[Any] = self._get_variance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise a__: str = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __a ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="epsilon" , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE = True , ) ->Union[DDPMSchedulerOutput, Tuple]: a__: Optional[Any] = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: a__ , a__: Optional[int] = torch.split(_SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: a__: str = None # 1. compute alphas, betas a__: List[Any] = self.alphas_cumprod[t] a__: List[Any] = self.alphas_cumprod[t - 1] if t > 0 else self.one a__: Optional[Any] = 1 - alpha_prod_t a__: int = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": a__: Any = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": a__: Optional[int] = model_output else: raise ValueError(F'Unsupported prediction_type {prediction_type}.' ) # 3. Clip "predicted x_0" a__: List[Any] = self.bit_scale if self.config.clip_sample: a__: str = torch.clamp(_SCREAMING_SNAKE_CASE , -scale , _SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf a__: Dict = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t a__: Optional[int] = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf a__: str = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise a__: str = 0 if t > 0: a__: int = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_SCREAMING_SNAKE_CASE ).to(model_output.device ) a__: Any = (self._get_variance(_SCREAMING_SNAKE_CASE , predicted_variance=_SCREAMING_SNAKE_CASE ) ** 0.5) * noise a__: Tuple = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase = 1.0 , ) -> str: '''simple docstring''' super().__init__() a__: List[Any] = bit_scale a__: str = ( ddim_bit_scheduler_step if isinstance(lowercase , lowercase) else ddpm_bit_scheduler_step ) self.register_modules(unet=lowercase , scheduler=lowercase) @torch.no_grad() def __call__( self , lowercase = 2_56 , lowercase = 2_56 , lowercase = 50 , lowercase = None , lowercase = 1 , lowercase = "pil" , lowercase = True , **lowercase , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' a__: Optional[Any] = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=lowercase , ) a__: Union[str, Any] = decimal_to_bits(lowercase) * self.bit_scale a__: Union[str, Any] = latents.to(self.device) self.scheduler.set_timesteps(lowercase) for t in self.progress_bar(self.scheduler.timesteps): # predict the noise residual a__: int = self.unet(lowercase , lowercase).sample # compute the previous noisy sample x_t -> x_t-1 a__: List[str] = self.scheduler.step(lowercase , lowercase , lowercase).prev_sample a__: str = bits_to_decimal(lowercase) if output_type == "pil": a__: int = self.numpy_to_pil(lowercase) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase)
290
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: Optional[int] = SamImageProcessor() a__: Tuple = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> List[Any]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[Any] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Optional[int] = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: List[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = self.get_image_processor() a__: List[Any] = SamProcessor(image_processor=lowercase) a__: Optional[int] = self.prepare_image_inputs() a__: Optional[Any] = image_processor(lowercase , return_tensors='np') a__: Tuple = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_torch def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.get_image_processor() a__: List[str] = SamProcessor(image_processor=lowercase) a__: Optional[Any] = [torch.ones((1, 3, 5, 5))] a__: Union[str, Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: int = processor.post_process_masks(lowercase , lowercase , lowercase) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Optional[int] = processor.post_process_masks( lowercase , torch.tensor(lowercase) , torch.tensor(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Dict = [np.ones((1, 3, 5, 5))] a__: Tuple = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = [[1, 0], [0, 1]] with self.assertRaises(lowercase): a__: List[Any] = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) @require_vision @require_tf class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: List[Any] = SamImageProcessor() a__: Optional[int] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> int: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[int] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Dict = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = self.get_image_processor() a__: str = SamProcessor(image_processor=lowercase) a__: int = self.prepare_image_inputs() a__: int = image_processor(lowercase , return_tensors='np') a__: Dict = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_tf def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Any = SamProcessor(image_processor=lowercase) a__: str = [tf.ones((1, 3, 5, 5))] a__: List[Any] = [[17_64, 26_46]] a__: List[Any] = [[6_83, 10_24]] a__: List[Any] = processor.post_process_masks(lowercase , lowercase , lowercase , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = processor.post_process_masks( lowercase , tf.convert_to_tensor(lowercase) , tf.convert_to_tensor(lowercase) , return_tensors='tf' , ) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Optional[Any] = [np.ones((1, 3, 5, 5))] a__: int = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: List[str] = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError): a__: Any = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = tempfile.mkdtemp() a__: int = SamImageProcessor() a__: Union[str, Any] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Any = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.get_image_processor() a__: int = SamProcessor(image_processor=lowercase) a__: int = np.random.randint(0 , 2 , size=(1, 3, 5, 5)).astype(np.floataa) a__: Dict = [tf.convert_to_tensor(lowercase)] a__: Union[str, Any] = [torch.tensor(lowercase)] a__: List[Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: Tuple = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='tf') a__: str = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='pt') self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy())) @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Dict = SamProcessor(image_processor=lowercase) a__: Any = self.prepare_image_inputs() a__: List[Any] = image_processor(lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Tuple = processor(images=lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Any = image_processor(lowercase , return_tensors='tf')['pixel_values'].numpy() a__: Any = processor(images=lowercase , return_tensors='tf')['pixel_values'].numpy() self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase))
290
1
"""simple docstring""" lowercase__ = {'a': ['c', 'b'], 'b': ['d', 'e'], 'c': [], 'd': [], 'e': []} lowercase__ = ['a', 'b', 'c', 'd', 'e'] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: List[Any] = start # add current to visited visited.append(_SCREAMING_SNAKE_CASE ) a__: Optional[int] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: a__: str = topological_sort(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # if all neighbors visited add current to sort sort.append(_SCREAMING_SNAKE_CASE ) # if all vertices haven't been visited select a new one to visit if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): for vertice in vertices: if vertice not in visited: a__: Optional[Any] = topological_sort(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # return sort return sort if __name__ == "__main__": lowercase__ = topological_sort('a', [], []) print(sort)
290
"""simple docstring""" from math import pow, sqrt def __a ( *_SCREAMING_SNAKE_CASE ) ->bool: a__: Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) > 0 and all(value > 0.0 for value in values ) return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """audio-spectrogram-transformer""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=16 , lowercase=True , lowercase=10 , lowercase=10 , lowercase=10_24 , lowercase=1_28 , **lowercase , ) -> str: '''simple docstring''' super().__init__(**lowercase) a__: Any = hidden_size a__: int = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Any = intermediate_size a__: Union[str, Any] = hidden_act a__: int = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: str = initializer_range a__: Tuple = layer_norm_eps a__: Any = patch_size a__: int = qkv_bias a__: Optional[Any] = frequency_stride a__: int = time_stride a__: List[str] = max_length a__: Tuple = num_mel_bins
290
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """roberta-prelayernorm""" def __init__( self , lowercase=5_02_65 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=True , lowercase=None , **lowercase , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase) a__: Union[str, Any] = vocab_size a__: str = hidden_size a__: Tuple = num_hidden_layers a__: List[str] = num_attention_heads a__: Dict = hidden_act a__: int = intermediate_size a__: Tuple = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: Tuple = max_position_embeddings a__: Tuple = type_vocab_size a__: Optional[Any] = initializer_range a__: Tuple = layer_norm_eps a__: Optional[int] = position_embedding_type a__: Any = use_cache a__: Dict = classifier_dropout class __snake_case ( __lowerCAmelCase ): @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": a__: str = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a__: Union[str, Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { 'configuration_blenderbot_small': [ 'BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlenderbotSmallConfig', 'BlenderbotSmallOnnxConfig', ], 'tokenization_blenderbot_small': ['BlenderbotSmallTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['BlenderbotSmallTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlenderbotSmallForCausalLM', 'BlenderbotSmallForConditionalGeneration', 'BlenderbotSmallModel', 'BlenderbotSmallPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'TFBlenderbotSmallForConditionalGeneration', 'TFBlenderbotSmallModel', 'TFBlenderbotSmallPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'FlaxBlenderbotSmallForConditionalGeneration', 'FlaxBlenderbotSmallModel', 'FlaxBlenderbotSmallPreTrainedModel', ] if TYPE_CHECKING: from .configuration_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_CONFIG_ARCHIVE_MAP, BlenderbotSmallConfig, BlenderbotSmallOnnxConfig, ) from .tokenization_blenderbot_small import BlenderbotSmallTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_blenderbot_small_fast import BlenderbotSmallTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blenderbot_small import ( BLENDERBOT_SMALL_PRETRAINED_MODEL_ARCHIVE_LIST, BlenderbotSmallForCausalLM, BlenderbotSmallForConditionalGeneration, BlenderbotSmallModel, BlenderbotSmallPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blenderbot_small import ( TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel, TFBlenderbotSmallPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, FlaxBlenderbotSmallPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """audio-spectrogram-transformer""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=16 , lowercase=True , lowercase=10 , lowercase=10 , lowercase=10_24 , lowercase=1_28 , **lowercase , ) -> str: '''simple docstring''' super().__init__(**lowercase) a__: Any = hidden_size a__: int = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Any = intermediate_size a__: Union[str, Any] = hidden_act a__: int = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: str = initializer_range a__: Tuple = layer_norm_eps a__: Any = patch_size a__: int = qkv_bias a__: Optional[Any] = frequency_stride a__: int = time_stride a__: List[str] = max_length a__: Tuple = num_mel_bins
290
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """roberta-prelayernorm""" def __init__( self , lowercase=5_02_65 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=True , lowercase=None , **lowercase , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase) a__: Union[str, Any] = vocab_size a__: str = hidden_size a__: Tuple = num_hidden_layers a__: List[str] = num_attention_heads a__: Dict = hidden_act a__: int = intermediate_size a__: Tuple = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: Tuple = max_position_embeddings a__: Tuple = type_vocab_size a__: Optional[Any] = initializer_range a__: Tuple = layer_norm_eps a__: Optional[int] = position_embedding_type a__: Any = use_cache a__: Dict = classifier_dropout class __snake_case ( __lowerCAmelCase ): @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": a__: str = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a__: Union[str, Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
290
"""simple docstring""" import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowercase__ = get_tests_dir('fixtures/test_sentencepiece.model') lowercase__ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') lowercase__ = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = CamembertTokenizer a__ = CamembertTokenizerFast a__ = True a__ = True def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing a__: Tuple = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Optional[Any] = '<pad>' a__: List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase) , lowercase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase) , lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: str = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>NOTUSED') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(lowercase) , 10_04) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_05) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[Any] = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) a__: List[Any] = CamembertTokenizerFast.from_pretrained(self.tmpdirname) a__: Dict = 'I was born in 92000, and this is falsé.' a__: Optional[int] = tokenizer.encode(lowercase) a__: Any = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) a__: Optional[Any] = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) a__: Tuple = tokenizer.convert_ids_to_tokens(lowercase) a__: Tuple = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' if not self.test_rust_tokenizer: return a__: Dict = self.get_tokenizer() a__: str = self.get_rust_tokenizer() a__: int = 'I was born in 92000, and this is falsé.' a__: Optional[Any] = tokenizer.tokenize(lowercase) a__: List[Any] = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) a__: str = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) a__: Tuple = self.get_rust_tokenizer() a__: Union[str, Any] = tokenizer.encode(lowercase) a__: List[Any] = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Union[str, Any] = {'input_ids': [[5, 54, 71_96, 2_97, 30, 23, 7_76, 18, 11, 32_15, 37_05, 82_52, 22, 31_64, 11_81, 21_16, 29, 16, 8_13, 25, 7_91, 33_14, 20, 34_46, 38, 2_75_75, 1_20, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 4_68, 17, 11, 90_88, 20, 15_17, 8, 2_28_04, 1_88_18, 10, 38, 6_29, 6_07, 6_07, 1_42, 19, 71_96, 8_67, 56, 1_03_26, 24, 22_67, 20, 4_16, 50_72, 1_56_12, 2_33, 7_34, 7, 23_99, 27, 16, 30_15, 16_49, 7, 24, 20, 43_38, 23_99, 27, 13, 34_00, 14, 13, 61_89, 8, 9_30, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. a__: int = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=lowercase , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=lowercase , )
290
1
"""simple docstring""" from statistics import mean import numpy as np def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->list: a__: List[str] = 0 # Number of processes finished a__: Optional[int] = 0 # Displays the finished process. # If it is 0, the performance is completed if it is 1, before the performance. a__: Union[str, Any] = [0] * no_of_process # List to include calculation results a__: Union[str, Any] = [0] * no_of_process # Sort by arrival time. a__: List[Any] = [burst_time[i] for i in np.argsort(_SCREAMING_SNAKE_CASE )] a__: Optional[int] = [process_name[i] for i in np.argsort(_SCREAMING_SNAKE_CASE )] arrival_time.sort() while no_of_process > finished_process_count: a__: Optional[int] = 0 while finished_process[i] == 1: i += 1 if current_time < arrival_time[i]: a__: List[Any] = arrival_time[i] a__: str = 0 # Index showing the location of the process being performed a__: Dict = 0 # Saves the current response ratio. a__: Optional[Any] = 0 for i in range(0 , _SCREAMING_SNAKE_CASE ): if finished_process[i] == 0 and arrival_time[i] <= current_time: a__: Tuple = (burst_time[i] + (current_time - arrival_time[i])) / burst_time[ i ] if response_ratio < temp: a__: Union[str, Any] = temp a__: Any = i # Calculate the turn around time a__: Dict = current_time + burst_time[loc] - arrival_time[loc] current_time += burst_time[loc] # Indicates that the process has been performed. a__: Any = 1 # Increase finished_process_count by 1 finished_process_count += 1 return turn_around_time def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->list: a__: Dict = [0] * no_of_process for i in range(0 , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = turn_around_time[i] - burst_time[i] return waiting_time if __name__ == "__main__": lowercase__ = 5 lowercase__ = ['A', 'B', 'C', 'D', 'E'] lowercase__ = [1, 2, 3, 4, 5] lowercase__ = [1, 2, 3, 4, 5] lowercase__ = calculate_turn_around_time( process_name, arrival_time, burst_time, no_of_process ) lowercase__ = calculate_waiting_time( process_name, turn_around_time, burst_time, no_of_process ) print('Process name \tArrival time \tBurst time \tTurn around time \tWaiting time') for i in range(0, no_of_process): print( f"{process_name[i]}\t\t{arrival_time[i]}\t\t{burst_time[i]}\t\t" f"{turn_around_time[i]}\t\t\t{waiting_time[i]}" ) print(f"average waiting time : {mean(waiting_time):.5f}") print(f"average turn around time : {mean(turn_around_time):.5f}")
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE = 1000000 ) ->int: a__: int = limit + 1 a__: Optional[int] = [0] * limit for first_term in range(1 , _SCREAMING_SNAKE_CASE ): for n in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a a__: Any = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f"{solution() = }")
290
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForMaskedImageModeling, HfArgumentParser, Trainer, TrainingArguments, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version lowercase__ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt') lowercase__ = list(MODEL_FOR_MASKED_IMAGE_MODELING_MAPPING.keys()) lowercase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __snake_case : a__ = field( default="""cifar10""" , metadata={"""help""": """Name of a dataset from the datasets package"""} ) a__ = field( default=__lowerCAmelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) a__ = field( default=__lowerCAmelCase , metadata={"""help""": """The column name of the images in the files. If not set, will try to use 'image' or 'img'."""} , ) a__ = field(default=__lowerCAmelCase , metadata={"""help""": """A folder containing the training data."""} ) a__ = field(default=__lowerCAmelCase , metadata={"""help""": """A folder containing the validation data."""} ) a__ = field( default=0.15 , metadata={"""help""": """Percent to split off of train for validation."""} ) a__ = field(default=32 , metadata={"""help""": """The size of the square patches to use for masking."""} ) a__ = field( default=0.6 , metadata={"""help""": """Percentage of patches to mask."""} , ) a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Optional[Any] = {} if self.train_dir is not None: a__: List[str] = self.train_dir if self.validation_dir is not None: a__: Dict = self.validation_dir a__: List[Any] = data_files if data_files else None @dataclass class __snake_case : a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """The model checkpoint for weights initialization. Can be a local path to a pytorch_model.bin or a """ """checkpoint identifier on the hub. """ """Don't set if you want to train a model from scratch.""" ) } , ) a__ = field( default=__lowerCAmelCase , metadata={"""help""": """If training from scratch, pass a model type from the list: """ + """, """.join(__lowerCAmelCase )} , ) a__ = field( default=__lowerCAmelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """Override some existing default config settings when a model is trained from scratch. Example: """ """n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index""" ) } , ) a__ = field( default=__lowerCAmelCase , metadata={"""help""": """Where do you want to store (cache) the pretrained models/datasets downloaded from the hub"""} , ) a__ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) a__ = field(default=__lowerCAmelCase , metadata={"""help""": """Name or path of preprocessor config."""} ) a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """The size (resolution) of each image. If not specified, will use `image_size` of the configuration.""" ) } , ) a__ = field( default=__lowerCAmelCase , metadata={ """help""": ( """The size (resolution) of each patch. If not specified, will use `patch_size` of the configuration.""" ) } , ) a__ = field( default=__lowerCAmelCase , metadata={"""help""": """Stride to use for the encoder."""} , ) class __snake_case : def __init__( self , lowercase=1_92 , lowercase=32 , lowercase=4 , lowercase=0.6) -> Optional[Any]: '''simple docstring''' a__: int = input_size a__: int = mask_patch_size a__: Dict = model_patch_size a__: int = mask_ratio if self.input_size % self.mask_patch_size != 0: raise ValueError('Input size must be divisible by mask patch size') if self.mask_patch_size % self.model_patch_size != 0: raise ValueError('Mask patch size must be divisible by model patch size') a__: Tuple = self.input_size // self.mask_patch_size a__: List[str] = self.mask_patch_size // self.model_patch_size a__: Dict = self.rand_size**2 a__: Any = int(np.ceil(self.token_count * self.mask_ratio)) def __call__( self) -> List[Any]: '''simple docstring''' a__: int = np.random.permutation(self.token_count)[: self.mask_count] a__: List[str] = np.zeros(self.token_count , dtype=lowercase) a__: Optional[int] = 1 a__: Optional[int] = mask.reshape((self.rand_size, self.rand_size)) a__: List[str] = mask.repeat(self.scale , axis=0).repeat(self.scale , axis=1) return torch.tensor(mask.flatten()) def __a ( _SCREAMING_SNAKE_CASE ) ->Dict: a__: str = torch.stack([example['pixel_values'] for example in examples] ) a__: str = torch.stack([example['mask'] for example in examples] ) return {"pixel_values": pixel_values, "bool_masked_pos": mask} def __a ( ) ->List[Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. a__: Dict = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. a__ , a__ , a__: Dict = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: a__ , a__ , a__: List[Any] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_mim' , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() a__: int = training_args.get_process_log_level() logger.setLevel(_SCREAMING_SNAKE_CASE ) transformers.utils.logging.set_verbosity(_SCREAMING_SNAKE_CASE ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + F'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(F'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. a__: Optional[int] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: a__: Any = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'Output directory ({training_args.output_dir}) already exists and is not empty. ' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Initialize our dataset. a__: int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. a__: Any = None if 'validation' in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , _SCREAMING_SNAKE_CASE ) and data_args.train_val_split > 0.0: a__: str = ds['train'].train_test_split(data_args.train_val_split ) a__: Any = split['train'] a__: Dict = split['test'] # Create config # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. a__: List[Any] = { 'cache_dir': model_args.cache_dir, 'revision': model_args.model_revision, 'use_auth_token': True if model_args.use_auth_token else None, } if model_args.config_name_or_path: a__: Any = AutoConfig.from_pretrained(model_args.config_name_or_path , **_SCREAMING_SNAKE_CASE ) elif model_args.model_name_or_path: a__: str = AutoConfig.from_pretrained(model_args.model_name_or_path , **_SCREAMING_SNAKE_CASE ) else: a__: int = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.config_overrides is not None: logger.info(F'Overriding config: {model_args.config_overrides}' ) config.update_from_string(model_args.config_overrides ) logger.info(F'New config: {config}' ) # make sure the decoder_type is "simmim" (only relevant for BEiT) if hasattr(_SCREAMING_SNAKE_CASE , 'decoder_type' ): a__: Union[str, Any] = 'simmim' # adapt config a__: Optional[int] = model_args.image_size if model_args.image_size is not None else config.image_size a__: Optional[Any] = model_args.patch_size if model_args.patch_size is not None else config.patch_size a__: List[str] = ( model_args.encoder_stride if model_args.encoder_stride is not None else config.encoder_stride ) config.update( { 'image_size': model_args.image_size, 'patch_size': model_args.patch_size, 'encoder_stride': model_args.encoder_stride, } ) # create image processor if model_args.image_processor_name: a__: Optional[int] = AutoImageProcessor.from_pretrained(model_args.image_processor_name , **_SCREAMING_SNAKE_CASE ) elif model_args.model_name_or_path: a__: Optional[int] = AutoImageProcessor.from_pretrained(model_args.model_name_or_path , **_SCREAMING_SNAKE_CASE ) else: a__: Optional[Any] = { conf.model_type: image_processor_class for conf, image_processor_class in IMAGE_PROCESSOR_MAPPING.items() } a__: Any = IMAGE_PROCESSOR_TYPES[model_args.model_type]() # create model if model_args.model_name_or_path: a__: Tuple = AutoModelForMaskedImageModeling.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info('Training new model from scratch' ) a__: Optional[Any] = AutoModelForMaskedImageModeling.from_config(_SCREAMING_SNAKE_CASE ) if training_args.do_train: a__: Any = ds['train'].column_names else: a__: List[Any] = ds['validation'].column_names if data_args.image_column_name is not None: a__: Optional[Any] = data_args.image_column_name elif "image" in column_names: a__: str = 'image' elif "img" in column_names: a__: List[str] = 'img' else: a__: Tuple = column_names[0] # transformations as done in original SimMIM paper # source: https://github.com/microsoft/SimMIM/blob/main/data/data_simmim.py a__: Union[str, Any] = Compose( [ Lambda(lambda _SCREAMING_SNAKE_CASE : img.convert('RGB' ) if img.mode != "RGB" else img ), RandomResizedCrop(model_args.image_size , scale=(0.67, 1.0) , ratio=(3.0 / 4.0, 4.0 / 3.0) ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) # create mask generator a__: int = MaskGenerator( input_size=model_args.image_size , mask_patch_size=data_args.mask_patch_size , model_patch_size=model_args.patch_size , mask_ratio=data_args.mask_ratio , ) def preprocess_images(_SCREAMING_SNAKE_CASE ): a__: str = [transforms(_SCREAMING_SNAKE_CASE ) for image in examples[image_column_name]] a__: Optional[int] = [mask_generator() for i in range(len(examples[image_column_name] ) )] return examples if training_args.do_train: if "train" not in ds: raise ValueError('--do_train requires a train dataset' ) if data_args.max_train_samples is not None: a__: Tuple = ds['train'].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(_SCREAMING_SNAKE_CASE ) if training_args.do_eval: if "validation" not in ds: raise ValueError('--do_eval requires a validation dataset' ) if data_args.max_eval_samples is not None: a__: Tuple = ( ds['validation'].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(_SCREAMING_SNAKE_CASE ) # Initialize our trainer a__: List[str] = Trainer( model=_SCREAMING_SNAKE_CASE , args=_SCREAMING_SNAKE_CASE , train_dataset=ds['train'] if training_args.do_train else None , eval_dataset=ds['validation'] if training_args.do_eval else None , tokenizer=_SCREAMING_SNAKE_CASE , data_collator=_SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: a__: Optional[int] = None if training_args.resume_from_checkpoint is not None: a__: List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: a__: List[str] = last_checkpoint a__: Optional[int] = trainer.train(resume_from_checkpoint=_SCREAMING_SNAKE_CASE ) trainer.save_model() trainer.log_metrics('train' , train_result.metrics ) trainer.save_metrics('train' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: a__: Optional[Any] = trainer.evaluate() trainer.log_metrics('eval' , _SCREAMING_SNAKE_CASE ) trainer.save_metrics('eval' , _SCREAMING_SNAKE_CASE ) # Write model card and (optionally) push to hub a__: Tuple = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'masked-image-modeling', 'dataset': data_args.dataset_name, 'tags': ['masked-image-modeling'], } if training_args.push_to_hub: trainer.push_to_hub(**_SCREAMING_SNAKE_CASE ) else: trainer.create_model_card(**_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
290
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union lowercase__ = TypeVar('T') lowercase__ = Union[List[T], Tuple[T, ...]] lowercase__ = Union[T, List[T], Dict[str, T]] lowercase__ = Union[str, bytes, os.PathLike]
290
1
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging lowercase__ = logging.get_logger(__name__) class __snake_case ( __lowerCAmelCase ): a__ = ["""input_features"""] def __init__( self , lowercase=80 , lowercase=1_60_00 , lowercase=1_60 , lowercase=30 , lowercase=4_00 , lowercase=0.0 , lowercase=False , **lowercase , ) -> List[Any]: '''simple docstring''' super().__init__( feature_size=lowercase , sampling_rate=lowercase , padding_value=lowercase , return_attention_mask=lowercase , **lowercase , ) a__: Union[str, Any] = n_fft a__: Dict = hop_length a__: Optional[int] = chunk_length a__: Any = chunk_length * sampling_rate a__: Union[str, Any] = self.n_samples // hop_length a__: Tuple = sampling_rate a__: int = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=lowercase , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=lowercase , norm='slaney' , mel_scale='slaney' , ) def lowerCamelCase_ ( self , lowercase) -> np.ndarray: '''simple docstring''' a__: Tuple = spectrogram( lowercase , window_function(self.n_fft , 'hann') , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='log10' , ) a__: Optional[int] = log_spec[:, :-1] a__: str = np.maximum(lowercase , log_spec.max() - 8.0) a__: List[str] = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowerCamelCase_ ( lowercase , lowercase , lowercase = 0.0) -> List[np.ndarray]: '''simple docstring''' if attention_mask is not None: a__: List[str] = np.array(lowercase , np.intaa) a__: List[Any] = [] for vector, length in zip(lowercase , attention_mask.sum(-1)): a__: Dict = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7) if length < normed_slice.shape[0]: a__: str = padding_value normed_input_values.append(lowercase) else: a__: List[str] = [(x - x.mean()) / np.sqrt(x.var() + 1e-7) for x in input_values] return normed_input_values def __call__( self , lowercase , lowercase = True , lowercase = None , lowercase = None , lowercase = None , lowercase = "max_length" , lowercase = None , lowercase = None , lowercase = None , **lowercase , ) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' f' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' f' was sampled with {self.sampling_rate} and not {sampling_rate}.') else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.') a__: Optional[int] = isinstance(lowercase , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(f'Only mono-channel audio is supported for input to {self}') a__: Any = is_batched_numpy or ( isinstance(lowercase , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: a__: List[str] = [np.asarray([speech] , dtype=np.floataa).T for speech in raw_speech] elif not is_batched and not isinstance(lowercase , np.ndarray): a__: Dict = np.asarray(lowercase , dtype=np.floataa) elif isinstance(lowercase , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): a__: Union[str, Any] = raw_speech.astype(np.floataa) # always return batch if not is_batched: a__: int = [np.asarray([raw_speech]).T] a__: Optional[Any] = BatchFeature({'input_features': raw_speech}) # convert into correct format for padding a__: Union[str, Any] = self.pad( lowercase , padding=lowercase , max_length=max_length if max_length else self.n_samples , truncation=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: a__: Any = self.zero_mean_unit_var_norm( padded_inputs['input_features'] , attention_mask=padded_inputs['attention_mask'] , padding_value=self.padding_value , ) a__: str = np.stack(padded_inputs['input_features'] , axis=0) # make sure list is in array format a__: Union[str, Any] = padded_inputs.get('input_features').transpose(2 , 0 , 1) a__: Dict = [self._np_extract_fbank_features(lowercase) for waveform in input_features[0]] if isinstance(input_features[0] , lowercase): a__: Optional[Any] = [np.asarray(lowercase , dtype=np.floataa) for feature in input_features] else: a__: Any = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) a__: Union[str, Any] = padded_inputs['attention_mask'][:, :: self.hop_length] if return_tensors is not None: a__: Dict = padded_inputs.convert_to_tensors(lowercase) return padded_inputs def lowerCamelCase_ ( self) -> Dict[str, Any]: '''simple docstring''' a__: Tuple = copy.deepcopy(self.__dict__) a__: Dict = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
290
"""simple docstring""" from math import pi, sqrt, tan def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) a__: List[Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(_SCREAMING_SNAKE_CASE , 2 ) * torus_radius * tube_radius def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) a__: int = (sidea + sidea + sidea) / 2 a__: Tuple = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(f"Rectangle: {area_rectangle(10, 20) = }") print(f"Square: {area_square(10) = }") print(f"Triangle: {area_triangle(10, 10) = }") print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(f"Parallelogram: {area_parallelogram(10, 20) = }") print(f"Rhombus: {area_rhombus(10, 20) = }") print(f"Trapezium: {area_trapezium(10, 20, 30) = }") print(f"Circle: {area_circle(20) = }") print(f"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(f"Cube: {surface_area_cube(20) = }") print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(f"Sphere: {surface_area_sphere(20) = }") print(f"Hemisphere: {surface_area_hemisphere(20) = }") print(f"Cone: {surface_area_cone(10, 20) = }") print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(f"Cylinder: {surface_area_cylinder(10, 20) = }") print(f"Torus: {surface_area_torus(20, 10) = }") print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(f"Square: {area_reg_polygon(4, 10) = }") print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
290
1
"""simple docstring""" lowercase__ = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: # Return True if there is node that has not iterated. a__: Optional[int] = [False] * len(_SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = [s] a__: int = True while queue: a__: int = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_SCREAMING_SNAKE_CASE ) a__: Optional[int] = True a__: Optional[Any] = u return visited[t] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Optional[int] = [-1] * (len(_SCREAMING_SNAKE_CASE )) a__: Tuple = 0 a__: int = [] a__: str = [i[:] for i in graph] # Record original cut, copy. while bfs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Tuple = float('Inf' ) a__: str = sink while s != source: # Find the minimum value in select path a__: str = min(_SCREAMING_SNAKE_CASE , graph[parent[s]][s] ) a__: Tuple = parent[s] max_flow += path_flow a__: List[str] = sink while v != source: a__: Optional[int] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow a__: str = parent[v] for i in range(len(_SCREAMING_SNAKE_CASE ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
290
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase__ = random.Random() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: if rng is None: a__: Any = global_rng a__: int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=4_00 , lowercase=20_00 , lowercase=1 , lowercase=0.0 , lowercase=1_60_00 , lowercase=True , lowercase=True , ) -> Union[str, Any]: '''simple docstring''' a__: Tuple = parent a__: Optional[int] = batch_size a__: Optional[Any] = min_seq_length a__: Optional[int] = max_seq_length a__: Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a__: Dict = feature_size a__: Any = padding_value a__: Optional[Any] = sampling_rate a__: Optional[Any] = return_attention_mask a__: str = do_normalize def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def lowerCamelCase_ ( self , lowercase=False , lowercase=False) -> Tuple: '''simple docstring''' def _flatten(lowercase): return list(itertools.chain(*lowercase)) if equal_length: a__: Dict = floats_list((self.batch_size, self.max_seq_length)) else: # make sure that inputs increase in size a__: List[Any] = [ _flatten(floats_list((x, self.feature_size))) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: a__: str = [np.asarray(lowercase) for x in speech_inputs] return speech_inputs class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = WavaVecaFeatureExtractor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = WavaVecaFeatureExtractionTester(self) def lowerCamelCase_ ( self , lowercase) -> List[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(lowercase , axis=0) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0) - 1) < 1e-3)) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 a__: Optional[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: List[str] = [np.asarray(lowercase) for speech_input in speech_inputs] # Test not batched input a__: Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='np').input_values a__: Dict = feat_extract(np_speech_inputs[0] , return_tensors='np').input_values self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test batched a__: Dict = feat_extract(lowercase , return_tensors='np').input_values a__: int = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test 2-D numpy arrays are batched. a__: int = [floats_list((1, x))[0] for x in (8_00, 8_00, 8_00)] a__: Union[str, Any] = np.asarray(lowercase) a__: int = feat_extract(lowercase , return_tensors='np').input_values a__: Any = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: List[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Optional[int] = ['longest', 'max_length', 'do_not_pad'] a__: List[Any] = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: Dict = feat_extract(lowercase , padding=lowercase , max_length=lowercase , return_tensors='np') a__: Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self.assertTrue(input_values[0][8_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self.assertTrue(input_values[0][10_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Optional[int] = range(8_00 , 14_00 , 2_00) a__: List[str] = [floats_list((1, x))[0] for x in lengths] a__: Tuple = ['longest', 'max_length', 'do_not_pad'] a__: Dict = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: int = feat_extract(lowercase , max_length=lowercase , padding=lowercase) a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Any = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Dict = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='max_length' , return_tensors='np') a__: int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1]) self._check_zero_mean_unit_variance(input_values[2]) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: int = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: str = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='longest' , return_tensors='np') a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00)) a__: Dict = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Tuple = feat_extract( lowercase , truncation=lowercase , max_length=20_00 , padding='longest' , return_tensors='np') a__: str = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00)) @require_torch def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' import torch a__: Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Tuple = np.random.rand(1_00).astype(np.floataa) a__: Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a__: Any = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np') self.assertTrue(np_processed.input_values.dtype == np.floataa) a__: Optional[Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt') self.assertTrue(pt_processed.input_values.dtype == torch.floataa) @slow @require_torch def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: a__: str = WavaVecaConfig.from_pretrained(lowercase) a__: str = WavaVecaFeatureExtractor.from_pretrained(lowercase) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer')
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) lowercase__ = { 'configuration_clip': [ 'CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CLIPConfig', 'CLIPOnnxConfig', 'CLIPTextConfig', 'CLIPVisionConfig', ], 'processing_clip': ['CLIPProcessor'], 'tokenization_clip': ['CLIPTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['CLIPTokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['CLIPFeatureExtractor'] lowercase__ = ['CLIPImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'CLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'CLIPModel', 'CLIPPreTrainedModel', 'CLIPTextModel', 'CLIPTextModelWithProjection', 'CLIPVisionModel', 'CLIPVisionModelWithProjection', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFCLIPModel', 'TFCLIPPreTrainedModel', 'TFCLIPTextModel', 'TFCLIPVisionModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'FlaxCLIPModel', 'FlaxCLIPPreTrainedModel', 'FlaxCLIPTextModel', 'FlaxCLIPTextPreTrainedModel', 'FlaxCLIPVisionModel', 'FlaxCLIPVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class __snake_case ( __lowerCAmelCase ): a__ = """decision_transformer""" a__ = ["""past_key_values"""] a__ = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=17 , lowercase=4 , lowercase=1_28 , lowercase=40_96 , lowercase=True , lowercase=1 , lowercase=10_24 , lowercase=3 , lowercase=1 , lowercase=None , lowercase="relu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=1e-5 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=5_02_56 , lowercase=5_02_56 , lowercase=False , lowercase=False , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = state_dim a__: int = act_dim a__: List[Any] = hidden_size a__: List[str] = max_ep_len a__: List[Any] = action_tanh a__: Optional[Any] = vocab_size a__: Tuple = n_positions a__: Dict = n_layer a__: Optional[int] = n_head a__: Optional[int] = n_inner a__: Any = activation_function a__: Union[str, Any] = resid_pdrop a__: Any = embd_pdrop a__: Any = attn_pdrop a__: List[Any] = layer_norm_epsilon a__: Optional[Any] = initializer_range a__: Any = scale_attn_weights a__: Dict = use_cache a__: Optional[int] = scale_attn_by_inverse_layer_idx a__: List[str] = reorder_and_upcast_attn a__: Any = bos_token_id a__: int = eos_token_id super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase)
290
1
"""simple docstring""" import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class __snake_case ( __lowerCAmelCase ): a__ = (CMStochasticIterativeScheduler,) a__ = 10 def lowerCamelCase_ ( self , **lowercase) -> Any: '''simple docstring''' a__: Tuple = { 'num_train_timesteps': 2_01, 'sigma_min': 0.002, 'sigma_max': 80.0, } config.update(**lowercase) return config def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = 10 a__: List[Any] = self.get_scheduler_config() a__: Optional[int] = self.scheduler_classes[0](**lowercase) scheduler.set_timesteps(lowercase) a__: Union[str, Any] = scheduler.timesteps[0] a__: Dict = scheduler.timesteps[1] a__: Union[str, Any] = self.dummy_sample a__: Optional[int] = 0.1 * sample a__: Optional[int] = scheduler.step(lowercase , lowercase , lowercase).prev_sample a__: int = scheduler.step(lowercase , lowercase , lowercase).prev_sample self.assertEqual(output_a.shape , sample.shape) self.assertEqual(output_a.shape , output_a.shape) def lowerCamelCase_ ( self) -> int: '''simple docstring''' for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowercase) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=lowercase) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Any = self.scheduler_classes[0] a__: str = self.get_scheduler_config() a__: str = scheduler_class(**lowercase) a__: Optional[int] = 1 scheduler.set_timesteps(lowercase) a__: List[str] = scheduler.timesteps a__: List[str] = torch.manual_seed(0) a__: int = self.dummy_model() a__: str = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(lowercase): # 1. scale model input a__: int = scheduler.scale_model_input(lowercase , lowercase) # 2. predict noise residual a__: List[str] = model(lowercase , lowercase) # 3. predict previous sample x_t-1 a__: Tuple = scheduler.step(lowercase , lowercase , lowercase , generator=lowercase).prev_sample a__: Optional[Any] = pred_prev_sample a__: Optional[int] = torch.sum(torch.abs(lowercase)) a__: List[Any] = torch.mean(torch.abs(lowercase)) assert abs(result_sum.item() - 192.7614) < 1e-2 assert abs(result_mean.item() - 0.2510) < 1e-3 def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: List[Any] = self.scheduler_classes[0] a__: List[Any] = self.get_scheduler_config() a__: List[Any] = scheduler_class(**lowercase) a__: Dict = [1_06, 0] scheduler.set_timesteps(timesteps=lowercase) a__: Union[str, Any] = scheduler.timesteps a__: Optional[int] = torch.manual_seed(0) a__: Tuple = self.dummy_model() a__: str = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input a__: Optional[Any] = scheduler.scale_model_input(lowercase , lowercase) # 2. predict noise residual a__: List[str] = model(lowercase , lowercase) # 3. predict previous sample x_t-1 a__: Union[str, Any] = scheduler.step(lowercase , lowercase , lowercase , generator=lowercase).prev_sample a__: Union[str, Any] = pred_prev_sample a__: List[str] = torch.sum(torch.abs(lowercase)) a__: Union[str, Any] = torch.mean(torch.abs(lowercase)) assert abs(result_sum.item() - 347.6357) < 1e-2 assert abs(result_mean.item() - 0.4527) < 1e-3 def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: int = self.scheduler_classes[0] a__: Any = self.get_scheduler_config() a__: Union[str, Any] = scheduler_class(**lowercase) a__: Tuple = [39, 30, 12, 15, 0] with self.assertRaises(lowercase , msg='`timesteps` must be in descending order.'): scheduler.set_timesteps(timesteps=lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Tuple = self.scheduler_classes[0] a__: Tuple = self.get_scheduler_config() a__: Any = scheduler_class(**lowercase) a__: Tuple = [39, 30, 12, 1, 0] a__: int = len(lowercase) with self.assertRaises(lowercase , msg='Can only pass one of `num_inference_steps` or `timesteps`.'): scheduler.set_timesteps(num_inference_steps=lowercase , timesteps=lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: str = self.scheduler_classes[0] a__: Union[str, Any] = self.get_scheduler_config() a__: int = scheduler_class(**lowercase) a__: Optional[int] = [scheduler.config.num_train_timesteps] with self.assertRaises( lowercase , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=lowercase)
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: while a != 0: a__ , a__: List[str] = b % a, a return b def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: if gcd(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) != 1: a__: Dict = F'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Union[str, Any] = 1, 0, a a__ , a__ , a__: Any = 0, 1, m while va != 0: a__: int = ua // va a__ , a__ , a__ , a__ , a__ , a__: Any = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
290
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor lowercase__ = logging.get_logger(__name__) class __snake_case ( __lowerCAmelCase ): def __init__( self , *lowercase , **lowercase) -> None: '''simple docstring''' warnings.warn( 'The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use OwlViTImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase)
290
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex lowercase__ = logging.getLogger(__name__) class __snake_case : def __init__( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' if not self.initialized: a__: Optional[int] = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Optional[int] = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.retriever.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ , a__: str = self.retriever._main_retrieve(lowercase , lowercase) return doc_ids, retrieved_doc_embeds class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> int: '''simple docstring''' if index is not None and index.is_initialized() and len(lowercase) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ') super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Any = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase) for worker in self.retrieval_workers ]) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' logger.info('initializing retrieval') if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if len(self.retrieval_workers) > 0: # Select a random retrieval actor. a__: int = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] a__ , a__: List[Any] = ray.get(random_worker.retrieve.remote(lowercase , lowercase)) else: a__ , a__: Dict = self._main_retrieve(lowercase , lowercase) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' return super(lowercase , cls).get_tokenizers(lowercase , lowercase , **lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = kwargs.pop('config' , lowercase) or RagConfig.from_pretrained(lowercase , **lowercase) a__: Union[str, Any] = RagTokenizer.from_pretrained(lowercase , config=lowercase) a__: int = rag_tokenizer.question_encoder a__: Any = rag_tokenizer.generator if indexed_dataset is not None: a__: List[Any] = 'custom' a__: Optional[Any] = CustomHFIndex(config.retrieval_vector_size , lowercase) else: a__: Dict = cls._build_index(lowercase) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
290
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=__lowerCAmelCase ): a__ = ["""onnx"""] def __init__( self , *lowercase , **lowercase) -> List[str]: '''simple docstring''' requires_backends(self , ['onnx']) @classmethod def lowerCamelCase_ ( cls , *lowercase , **lowercase) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['onnx']) @classmethod def lowerCamelCase_ ( cls , *lowercase , **lowercase) -> Optional[Any]: '''simple docstring''' requires_backends(cls , ['onnx'])
290
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: int = None if token is not None: a__: Tuple = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Optional[Any] = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: str = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) a__: int = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Dict = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Dict: a__: Dict = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Dict = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: List[Any] = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Optional[int] = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: List[Any] = None if token is not None: a__: Optional[int] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = result.headers['Location'] a__: Optional[int] = requests.get(_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: int = os.path.join(_SCREAMING_SNAKE_CASE , F'{artifact_name}.zip' ) with open(_SCREAMING_SNAKE_CASE , 'wb' ) as fp: fp.write(response.content ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: a__: List[Any] = [] a__: Optional[Any] = [] a__: List[Any] = None with zipfile.ZipFile(_SCREAMING_SNAKE_CASE ) as z: for filename in z.namelist(): if not os.path.isdir(_SCREAMING_SNAKE_CASE ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(_SCREAMING_SNAKE_CASE ) as f: for line in f: a__: Optional[int] = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs a__: Union[str, Any] = line[: line.index(': ' )] a__: Union[str, Any] = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed a__: Optional[int] = line[len('FAILED ' ) :] failed_tests.append(_SCREAMING_SNAKE_CASE ) elif filename == "job_name.txt": a__: Union[str, Any] = line if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( F'`errors` and `failed_tests` should have the same number of elements. Got {len(_SCREAMING_SNAKE_CASE )} for `errors` ' F'and {len(_SCREAMING_SNAKE_CASE )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some' ' problem.' ) a__: Tuple = None if job_name and job_links: a__: Dict = job_links.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # A list with elements of the form (line of error, error, failed test) a__: int = [x + [y] + [job_link] for x, y in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )] return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->str: a__: int = [] a__: Optional[int] = [os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for p in os.listdir(_SCREAMING_SNAKE_CASE ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(_SCREAMING_SNAKE_CASE , job_links=_SCREAMING_SNAKE_CASE ) ) return errors def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Any: a__: str = Counter() counter.update([x[1] for x in logs] ) a__: int = counter.most_common() a__: Any = {} for error, count in counts: if error_filter is None or error not in error_filter: a__: List[str] = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} a__: Optional[Any] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: List[str] = test.split('::' )[0] if test.startswith('tests/models/' ): a__: Dict = test.split('/' )[2] else: a__: Any = None return test def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: int = [(x[0], x[1], get_model(x[2] )) for x in logs] a__: List[Any] = [x for x in logs if x[2] is not None] a__: Optional[Any] = {x[2] for x in logs} a__: Dict = {} for test in tests: a__: Union[str, Any] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) a__: Union[str, Any] = counter.most_common() a__: List[str] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} a__: List[Any] = sum(error_counts.values() ) if n_errors > 0: a__: Any = {'count': n_errors, 'errors': error_counts} a__: Optional[int] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Any = '| no. | error | status |' a__: Any = '|-:|:-|:-|' a__: str = [header, sep] for error in reduced_by_error: a__: int = reduced_by_error[error]['count'] a__: Tuple = F'| {count} | {error[:100]} | |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->str: a__: List[str] = '| model | no. of errors | major error | count |' a__: str = '|-:|-:|-:|-:|' a__: int = [header, sep] for model in reduced_by_model: a__: Tuple = reduced_by_model[model]['count'] a__ , a__: Dict = list(reduced_by_model[model]['errors'].items() )[0] a__: Dict = F'| {model} | {count} | {error[:60]} | {_count} |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') lowercase__ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowercase__ = get_job_links(args.workflow_run_id, token=args.token) lowercase__ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowercase__ = k.find(' / ') lowercase__ = k[index + len(' / ') :] lowercase__ = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowercase__ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowercase__ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowercase__ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowercase__ = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowercase__ = reduce_by_error(errors) lowercase__ = reduce_by_model(errors) lowercase__ = make_github_table(reduced_by_error) lowercase__ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
290
1
"""simple docstring""" import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class __snake_case : def __init__( self , lowercase , lowercase=14 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=4 , lowercase=4 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=0.02 , ) -> str: '''simple docstring''' a__: Optional[int] = parent a__: List[str] = batch_size a__: List[Any] = seq_length a__: Any = is_training a__: Dict = use_input_mask a__: int = use_token_type_ids a__: List[Any] = use_labels a__: Tuple = vocab_size a__: List[Any] = hidden_size a__: List[Any] = rotary_dim a__: List[Any] = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Tuple = intermediate_size a__: Union[str, Any] = hidden_act a__: List[str] = hidden_dropout_prob a__: int = attention_probs_dropout_prob a__: List[Any] = max_position_embeddings a__: Any = initializer_range a__: str = None a__: Dict = vocab_size - 1 a__: List[str] = vocab_size - 1 a__: Optional[Any] = vocab_size - 1 def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a__: Union[str, Any] = None if self.use_input_mask: a__: Any = random_attention_mask([self.batch_size, self.seq_length]) a__: Tuple = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=lowercase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: int = self.prepare_config_and_inputs() a__ , a__ , a__: Tuple = config_and_inputs a__: Tuple = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> Optional[int]: '''simple docstring''' a__: List[str] = 20 a__: Dict = model_class_name(lowercase) a__: Optional[Any] = model.init_cache(input_ids.shape[0] , lowercase) a__: Any = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype='i4') a__: Optional[Any] = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1)[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1)) a__: Tuple = model( input_ids[:, :-1] , attention_mask=lowercase , past_key_values=lowercase , position_ids=lowercase , ) a__: Optional[int] = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='i4') a__: Optional[int] = model( input_ids[:, -1:] , attention_mask=lowercase , past_key_values=outputs_cache.past_key_values , position_ids=lowercase , ) a__: List[Any] = model(lowercase) a__: Optional[int] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1e-3 , msg=f'Max diff is {diff}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> Any: '''simple docstring''' a__: str = 20 a__: int = model_class_name(lowercase) a__: Optional[Any] = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]))] , axis=-1 , ) a__: Tuple = model.init_cache(input_ids.shape[0] , lowercase) a__: str = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1)[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1)) a__: Dict = model( input_ids[:, :-1] , attention_mask=lowercase , past_key_values=lowercase , position_ids=lowercase , ) a__: Any = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='i4') a__: Optional[int] = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=lowercase , position_ids=lowercase , ) a__: Optional[int] = model(lowercase , attention_mask=lowercase) a__: int = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1e-3 , msg=f'Max diff is {diff}') @require_flax class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () a__ = (FlaxGPTJForCausalLM,) if is_flax_available() else () def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = FlaxGPTJModelTester(self) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' for model_class_name in self.all_model_classes: a__ , a__ , a__: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(lowercase , lowercase , lowercase , lowercase) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' for model_class_name in self.all_model_classes: a__ , a__ , a__: Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( lowercase , lowercase , lowercase , lowercase) @tooslow def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[int] = GPTaTokenizer.from_pretrained('gpt2' , pad_token='<|endoftext|>' , padding_side='left') a__: str = tokenizer(['Hello this is a long string', 'Hey'] , return_tensors='np' , padding=lowercase , truncation=lowercase) a__: Union[str, Any] = FlaxGPTJForCausalLM.from_pretrained('EleutherAI/gpt-j-6B') a__: Tuple = False a__: Optional[Any] = model.config.eos_token_id a__: Optional[Any] = jax.jit(model.generate) a__: Optional[int] = jit_generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , pad_token_id=tokenizer.pad_token_id).sequences a__: List[str] = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase) a__: List[str] = [ 'Hello this is a long string of text.\n\nI\'m trying to get the text of the', 'Hey, I\'m a little late to the party. I\'m going to', ] self.assertListEqual(lowercase , lowercase) @is_pt_flax_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__ , a__: Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): # prepare inputs a__: Union[str, Any] = self._prepare_for_class(lowercase , lowercase) a__: List[str] = {k: torch.tensor(v.tolist()) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class a__: Tuple = model_class.__name__[4:] # Skip the "Flax" at the beginning a__: Union[str, Any] = getattr(lowercase , lowercase) a__ , a__: Tuple = pt_inputs['input_ids'].shape a__: List[Any] = np.random.randint(0 , seq_length - 1 , size=(batch_size,)) for batch_idx, start_index in enumerate(lowercase): a__: Union[str, Any] = 0 a__: Tuple = 1 a__: Tuple = 0 a__: Dict = 1 a__: str = pt_model_class(lowercase).eval() a__: int = model_class(lowercase , dtype=jnp.floataa) a__: Union[str, Any] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , lowercase) a__: int = fx_state with torch.no_grad(): a__: Tuple = pt_model(**lowercase).to_tuple() a__: Optional[int] = fx_model(**lowercase).to_tuple() self.assertEqual(len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowercase) a__: Optional[Any] = model_class.from_pretrained(lowercase , from_pt=lowercase) a__: Optional[int] = fx_model_loaded(**lowercase).to_tuple() self.assertEqual( len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output_loaded, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4e-2) @is_pt_flax_cross_test def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__ , a__: Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): # prepare inputs a__: int = self._prepare_for_class(lowercase , lowercase) a__: Dict = {k: torch.tensor(v.tolist()) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class a__: List[str] = model_class.__name__[4:] # Skip the "Flax" at the beginning a__: List[Any] = getattr(lowercase , lowercase) a__: str = pt_model_class(lowercase).eval() a__: List[str] = model_class(lowercase , dtype=jnp.floataa) a__: List[str] = load_flax_weights_in_pytorch_model(lowercase , fx_model.params) a__ , a__: Tuple = pt_inputs['input_ids'].shape a__: List[Any] = np.random.randint(0 , seq_length - 1 , size=(batch_size,)) for batch_idx, start_index in enumerate(lowercase): a__: Any = 0 a__: Tuple = 1 a__: Union[str, Any] = 0 a__: Union[str, Any] = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): a__: Optional[Any] = pt_model(**lowercase).to_tuple() a__: List[Any] = fx_model(**lowercase).to_tuple() self.assertEqual(len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowercase) a__: Dict = pt_model_class.from_pretrained(lowercase , from_flax=lowercase) with torch.no_grad(): a__: str = pt_model_loaded(**lowercase).to_tuple() self.assertEqual( len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2) @tooslow def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' for model_class_name in self.all_model_classes: a__: int = model_class_name.from_pretrained('EleutherAI/gpt-j-6B') a__: Tuple = model(np.ones((1, 1))) self.assertIsNotNone(lowercase)
290
"""simple docstring""" import math def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __a ( _SCREAMING_SNAKE_CASE = 0.1 ) ->int: a__: str = 3 a__: Optional[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_SCREAMING_SNAKE_CASE ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" lowercase__ = 256 # Modulus to hash a string lowercase__ = 1000003 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->bool: a__: str = len(_SCREAMING_SNAKE_CASE ) a__: Dict = len(_SCREAMING_SNAKE_CASE ) if p_len > t_len: return False a__: List[str] = 0 a__: List[Any] = 0 a__: int = 1 # Calculating the hash of pattern and substring of text for i in range(_SCREAMING_SNAKE_CASE ): a__: List[str] = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus a__: int = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue a__: Optional[Any] = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash a__: Any = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def __a ( ) ->None: a__: Optional[int] = 'abc1abc12' a__: Optional[int] = 'alskfjaldsabc1abc1abc12k23adsfabcabc' a__: Tuple = 'alskfjaldsk23adsfabcabc' assert rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and not rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Test 2) a__: Dict = 'ABABX' a__: int = 'ABABZABABYABABX' assert rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Test 3) a__: str = 'AAAB' a__: Tuple = 'ABAAAAAB' assert rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Test 4) a__: Dict = 'abcdabcy' a__: Any = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Test 5) a__: int = 'Lü' a__: Dict = 'Lüsai' assert rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = 'Lue' assert not rabin_karp(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
290
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ = {'processing_wav2vec2_with_lm': ['Wav2Vec2ProcessorWithLM']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
1
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva lowercase__ = '' lowercase__ = '' lowercase__ = '' lowercase__ = 1 # (0 is vertical, 1 is horizontal) def __a ( ) ->None: a__ , a__: Dict = get_dataset(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print('Processing...' ) a__ , a__ , a__: List[Any] = update_image_and_anno(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for index, image in enumerate(_SCREAMING_SNAKE_CASE ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' a__: List[str] = random_chars(32 ) a__: List[str] = paths[index].split(os.sep )[-1].rsplit('.' , 1 )[0] a__: int = F'{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}' cva.imwrite(F'/{file_root}.jpg' , _SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F'Success {index+1}/{len(_SCREAMING_SNAKE_CASE )} with {file_name}' ) a__: Any = [] for anno in new_annos[index]: a__: Union[str, Any] = F'{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}' annos_list.append(_SCREAMING_SNAKE_CASE ) with open(F'/{file_root}.txt' , 'w' ) as outfile: outfile.write('\n'.join(line for line in annos_list ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->tuple[list, list]: a__: Optional[int] = [] a__: str = [] for label_file in glob.glob(os.path.join(_SCREAMING_SNAKE_CASE , '*.txt' ) ): a__: int = label_file.split(os.sep )[-1].rsplit('.' , 1 )[0] with open(_SCREAMING_SNAKE_CASE ) as in_file: a__: Dict = in_file.readlines() a__: Optional[Any] = os.path.join(_SCREAMING_SNAKE_CASE , F'{label_name}.jpg' ) a__: Union[str, Any] = [] for obj_list in obj_lists: a__: Optional[int] = obj_list.rstrip('\n' ).split(' ' ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(_SCREAMING_SNAKE_CASE ) labels.append(_SCREAMING_SNAKE_CASE ) return img_paths, labels def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 1 ) ->tuple[list, list, list]: a__: Optional[Any] = [] a__: str = [] a__: List[Any] = [] for idx in range(len(_SCREAMING_SNAKE_CASE ) ): a__: List[str] = [] a__: Tuple = img_list[idx] path_list.append(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = anno_list[idx] a__: Union[str, Any] = cva.imread(_SCREAMING_SNAKE_CASE ) if flip_type == 1: a__: Dict = cva.flip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for bbox in img_annos: a__: Optional[int] = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: a__: Dict = cva.flip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for bbox in img_annos: a__: Dict = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(_SCREAMING_SNAKE_CASE ) new_imgs_list.append(_SCREAMING_SNAKE_CASE ) return new_imgs_list, new_annos_lists, path_list def __a ( _SCREAMING_SNAKE_CASE = 32 ) ->str: assert number_char > 1, "The number of character should greater than 1" a__: Dict = ascii_lowercase + digits return "".join(random.choice(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print('DONE ✅')
290
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = KandinskyInpaintPipeline a__ = ["""prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] a__ = [ """prompt""", """negative_prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] a__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """negative_prompt""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] a__ = False @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 1_00 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[int] = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base') return tokenizer @property def lowerCamelCase_ ( self) -> Any: '''simple docstring''' torch.manual_seed(0) a__: Dict = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=10_05 , ) a__: Optional[Any] = MultilingualCLIP(lowercase) a__: int = text_encoder.eval() return text_encoder @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' torch.manual_seed(0) a__: Any = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } a__: str = UNetaDConditionModel(**lowercase) return model @property def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' torch.manual_seed(0) a__: Any = VQModel(**self.dummy_movq_kwargs) return model def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = self.dummy_text_encoder a__: int = self.dummy_tokenizer a__: str = self.dummy_unet a__: Any = self.dummy_movq a__: Tuple = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.00085 , beta_end=0.012 , clip_sample=lowercase , set_alpha_to_one=lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=lowercase , ) a__: Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> Any: '''simple docstring''' a__: List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase)).to(lowercase) a__: int = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1)).to(lowercase) # create init_image a__: Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) a__: int = image.cpu().permute(0 , 2 , 3 , 1)[0] a__: Optional[int] = Image.fromarray(np.uinta(lowercase)).convert('RGB').resize((2_56, 2_56)) # create mask a__: Tuple = np.ones((64, 64) , dtype=np.floataa) a__: Optional[Any] = 0 if str(lowercase).startswith('mps'): a__: str = torch.manual_seed(lowercase) else: a__: Dict = torch.Generator(device=lowercase).manual_seed(lowercase) a__: Optional[int] = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[Any] = 'cpu' a__: List[Any] = self.get_dummy_components() a__: Optional[Any] = self.pipeline_class(**lowercase) a__: str = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[int] = pipe(**self.get_dummy_inputs(lowercase)) a__: List[str] = output.images a__: int = pipe( **self.get_dummy_inputs(lowercase) , return_dict=lowercase , )[0] a__: Optional[Any] = image[0, -3:, -3:, -1] a__: List[Any] = image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}') assert image.shape == (1, 64, 64, 3) a__: str = np.array( [0.8326919, 0.73790467, 0.20918581, 0.9309612, 0.5511791, 0.43713328, 0.5513321, 0.49922934, 0.59497786]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy') a__: int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png') a__: Union[str, Any] = np.ones((7_68, 7_68) , dtype=np.floataa) a__: int = 0 a__: Optional[int] = 'a hat' a__: int = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa) pipe_prior.to(lowercase) a__: Any = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa) a__: Optional[Any] = pipeline.to(lowercase) pipeline.set_progress_bar_config(disable=lowercase) a__: Dict = torch.Generator(device='cpu').manual_seed(0) a__ , a__: Optional[Any] = pipe_prior( lowercase , generator=lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() a__: List[str] = pipeline( lowercase , image=lowercase , mask_image=lowercase , image_embeds=lowercase , negative_image_embeds=lowercase , generator=lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) a__: str = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(lowercase , lowercase)
290
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __snake_case : def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=32 , lowercase=2 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=False , lowercase=True , lowercase="None" , lowercase=3 , lowercase=4 , lowercase=None , ) -> Union[str, Any]: '''simple docstring''' a__: Union[str, Any] = parent a__: str = batch_size a__: List[str] = seq_length a__: int = is_training a__: Union[str, Any] = use_input_mask a__: str = use_token_type_ids a__: Union[str, Any] = use_labels a__: Any = vocab_size a__: str = hidden_size a__: Tuple = num_hidden_layers a__: Optional[Any] = num_attention_heads a__: Dict = intermediate_size a__: Tuple = hidden_act a__: List[Any] = hidden_dropout_prob a__: Tuple = attention_probs_dropout_prob a__: int = max_position_embeddings a__: int = type_vocab_size a__: Any = type_sequence_label_size a__: Optional[int] = initializer_range a__: List[Any] = num_labels a__: Optional[Any] = num_choices a__: Union[str, Any] = relative_attention a__: Optional[int] = position_biased_input a__: Tuple = pos_att_type a__: Union[str, Any] = scope def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a__: str = None if self.use_input_mask: a__: List[str] = random_attention_mask([self.batch_size, self.seq_length]) a__: List[Any] = None if self.use_token_type_ids: a__: Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) a__: str = None a__: List[str] = None a__: List[Any] = None if self.use_labels: a__: List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size) a__: Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) a__: List[Any] = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=lowercase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = TFDebertaVaModel(config=lowercase) a__: Dict = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} a__: Optional[Any] = [input_ids, input_mask] a__: Optional[Any] = model(lowercase) a__: int = model(lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Optional[Any] = TFDebertaVaForMaskedLM(config=lowercase) a__: Optional[int] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a__: Any = model(lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) -> Optional[int]: '''simple docstring''' a__: Dict = self.num_labels a__: Union[str, Any] = TFDebertaVaForSequenceClassification(config=lowercase) a__: Optional[int] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a__: List[Any] = model(lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.num_labels a__: int = TFDebertaVaForTokenClassification(config=lowercase) a__: Optional[int] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a__: Any = model(lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' a__: Dict = TFDebertaVaForQuestionAnswering(config=lowercase) a__: str = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a__: List[str] = model(lowercase) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: List[Any] = self.prepare_config_and_inputs() ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ): Tuple = config_and_inputs a__: Optional[int] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) a__ = ( { """feature-extraction""": TFDebertaVaModel, """fill-mask""": TFDebertaVaForMaskedLM, """question-answering""": TFDebertaVaForQuestionAnswering, """text-classification""": TFDebertaVaForSequenceClassification, """token-classification""": TFDebertaVaForTokenClassification, """zero-shot""": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) a__ = False a__ = False def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: str = TFDebertaVaModelTester(self) a__: Any = ConfigTester(self , config_class=lowercase , hidden_size=37) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase) @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[int] = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge') self.assertIsNotNone(lowercase) @require_tf class __snake_case ( unittest.TestCase ): @unittest.skip(reason='Model not available yet') def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' pass @slow def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Tuple = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge') a__: Optional[int] = tf.constant([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]]) a__: Any = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) a__: Tuple = model(lowercase , attention_mask=lowercase)[0] a__: Optional[int] = tf.constant( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]]) tf.debugging.assert_near(output[:, 1:4, 1:4] , lowercase , atol=1e-4)
290
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() lowercase__ = logging.get_logger('transformers.models.encodec') lowercase__ = { 'quantizer.vq.layers.*._codebook.inited': 'quantizer.layers.*.codebook.inited', 'quantizer.vq.layers.*._codebook.cluster_size': 'quantizer.layers.*.codebook.cluster_size', 'quantizer.vq.layers.*._codebook.embed': 'quantizer.layers.*.codebook.embed', 'quantizer.vq.layers.*._codebook.embed_avg': 'quantizer.layers.*.codebook.embed_avg', } lowercase__ = { 'encoder.model.0.conv.conv': 'encoder.layers.0.conv', 'encoder.model.1.block.1.conv.conv': 'encoder.layers.1.block.1.conv', 'encoder.model.1.block.3.conv.conv': 'encoder.layers.1.block.3.conv', 'encoder.model.1.shortcut.conv.conv': 'encoder.layers.1.shortcut.conv', 'encoder.model.3.conv.conv': 'encoder.layers.3.conv', 'encoder.model.4.block.1.conv.conv': 'encoder.layers.4.block.1.conv', 'encoder.model.4.block.3.conv.conv': 'encoder.layers.4.block.3.conv', 'encoder.model.4.shortcut.conv.conv': 'encoder.layers.4.shortcut.conv', 'encoder.model.6.conv.conv': 'encoder.layers.6.conv', 'encoder.model.7.block.1.conv.conv': 'encoder.layers.7.block.1.conv', 'encoder.model.7.block.3.conv.conv': 'encoder.layers.7.block.3.conv', 'encoder.model.7.shortcut.conv.conv': 'encoder.layers.7.shortcut.conv', 'encoder.model.9.conv.conv': 'encoder.layers.9.conv', 'encoder.model.10.block.1.conv.conv': 'encoder.layers.10.block.1.conv', 'encoder.model.10.block.3.conv.conv': 'encoder.layers.10.block.3.conv', 'encoder.model.10.shortcut.conv.conv': 'encoder.layers.10.shortcut.conv', 'encoder.model.12.conv.conv': 'encoder.layers.12.conv', 'encoder.model.13.lstm': 'encoder.layers.13.lstm', 'encoder.model.15.conv.conv': 'encoder.layers.15.conv', } lowercase__ = { 'encoder.model.0.conv.norm': 'encoder.layers.0.norm', 'encoder.model.1.block.1.conv.norm': 'encoder.layers.1.block.1.norm', 'encoder.model.1.block.3.conv.norm': 'encoder.layers.1.block.3.norm', 'encoder.model.1.shortcut.conv.norm': 'encoder.layers.1.shortcut.norm', 'encoder.model.3.conv.norm': 'encoder.layers.3.norm', 'encoder.model.4.block.1.conv.norm': 'encoder.layers.4.block.1.norm', 'encoder.model.4.block.3.conv.norm': 'encoder.layers.4.block.3.norm', 'encoder.model.4.shortcut.conv.norm': 'encoder.layers.4.shortcut.norm', 'encoder.model.6.conv.norm': 'encoder.layers.6.norm', 'encoder.model.7.block.1.conv.norm': 'encoder.layers.7.block.1.norm', 'encoder.model.7.block.3.conv.norm': 'encoder.layers.7.block.3.norm', 'encoder.model.7.shortcut.conv.norm': 'encoder.layers.7.shortcut.norm', 'encoder.model.9.conv.norm': 'encoder.layers.9.norm', 'encoder.model.10.block.1.conv.norm': 'encoder.layers.10.block.1.norm', 'encoder.model.10.block.3.conv.norm': 'encoder.layers.10.block.3.norm', 'encoder.model.10.shortcut.conv.norm': 'encoder.layers.10.shortcut.norm', 'encoder.model.12.conv.norm': 'encoder.layers.12.norm', 'encoder.model.15.conv.norm': 'encoder.layers.15.norm', } lowercase__ = { 'decoder.model.0.conv.conv': 'decoder.layers.0.conv', 'decoder.model.1.lstm': 'decoder.layers.1.lstm', 'decoder.model.3.convtr.convtr': 'decoder.layers.3.conv', 'decoder.model.4.block.1.conv.conv': 'decoder.layers.4.block.1.conv', 'decoder.model.4.block.3.conv.conv': 'decoder.layers.4.block.3.conv', 'decoder.model.4.shortcut.conv.conv': 'decoder.layers.4.shortcut.conv', 'decoder.model.6.convtr.convtr': 'decoder.layers.6.conv', 'decoder.model.7.block.1.conv.conv': 'decoder.layers.7.block.1.conv', 'decoder.model.7.block.3.conv.conv': 'decoder.layers.7.block.3.conv', 'decoder.model.7.shortcut.conv.conv': 'decoder.layers.7.shortcut.conv', 'decoder.model.9.convtr.convtr': 'decoder.layers.9.conv', 'decoder.model.10.block.1.conv.conv': 'decoder.layers.10.block.1.conv', 'decoder.model.10.block.3.conv.conv': 'decoder.layers.10.block.3.conv', 'decoder.model.10.shortcut.conv.conv': 'decoder.layers.10.shortcut.conv', 'decoder.model.12.convtr.convtr': 'decoder.layers.12.conv', 'decoder.model.13.block.1.conv.conv': 'decoder.layers.13.block.1.conv', 'decoder.model.13.block.3.conv.conv': 'decoder.layers.13.block.3.conv', 'decoder.model.13.shortcut.conv.conv': 'decoder.layers.13.shortcut.conv', 'decoder.model.15.conv.conv': 'decoder.layers.15.conv', } lowercase__ = { 'decoder.model.0.conv.norm': 'decoder.layers.0.norm', 'decoder.model.3.convtr.norm': 'decoder.layers.3.norm', 'decoder.model.4.block.1.conv.norm': 'decoder.layers.4.block.1.norm', 'decoder.model.4.block.3.conv.norm': 'decoder.layers.4.block.3.norm', 'decoder.model.4.shortcut.conv.norm': 'decoder.layers.4.shortcut.norm', 'decoder.model.6.convtr.norm': 'decoder.layers.6.norm', 'decoder.model.7.block.1.conv.norm': 'decoder.layers.7.block.1.norm', 'decoder.model.7.block.3.conv.norm': 'decoder.layers.7.block.3.norm', 'decoder.model.7.shortcut.conv.norm': 'decoder.layers.7.shortcut.norm', 'decoder.model.9.convtr.norm': 'decoder.layers.9.norm', 'decoder.model.10.block.1.conv.norm': 'decoder.layers.10.block.1.norm', 'decoder.model.10.block.3.conv.norm': 'decoder.layers.10.block.3.norm', 'decoder.model.10.shortcut.conv.norm': 'decoder.layers.10.shortcut.norm', 'decoder.model.12.convtr.norm': 'decoder.layers.12.norm', 'decoder.model.13.block.1.conv.norm': 'decoder.layers.13.block.1.norm', 'decoder.model.13.block.3.conv.norm': 'decoder.layers.13.block.3.norm', 'decoder.model.13.shortcut.conv.norm': 'decoder.layers.13.shortcut.norm', 'decoder.model.15.conv.norm': 'decoder.layers.15.norm', } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } lowercase__ = [] lowercase__ = [] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: for attribute in key.split('.' ): a__: str = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: str = value elif weight_type == "weight_g": a__: int = value elif weight_type == "weight_v": a__: Tuple = value elif weight_type == "bias": a__: Dict = value elif weight_type == "running_mean": a__: Any = value elif weight_type == "running_var": a__: Tuple = value elif weight_type == "num_batches_tracked": a__: List[str] = value elif weight_type == "weight_ih_l0": a__: List[Any] = value elif weight_type == "weight_hh_l0": a__: List[Any] = value elif weight_type == "bias_ih_l0": a__: List[Any] = value elif weight_type == "bias_hh_l0": a__: List[Any] = value elif weight_type == "weight_ih_l1": a__: int = value elif weight_type == "weight_hh_l1": a__: str = value elif weight_type == "bias_ih_l1": a__: Union[str, Any] = value elif weight_type == "bias_hh_l1": a__: Any = value else: a__: Union[str, Any] = value logger.info(F'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: for key in ignore_keys: if key.endswith('.*' ): if name.startswith(key[:-1] ): return True elif ".*." in key: a__ , a__: Optional[Any] = key.split('.*.' ) if prefix in name and suffix in name: return True elif key in name: return True return False def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[Any] = [] if model_name == "encodec_24khz" or "encodec_32khz": a__: Optional[int] = MAPPING_24K elif model_name == "encodec_48khz": a__: List[Any] = MAPPING_48K else: raise ValueError(F'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): logger.info(F'{name} was ignored' ) continue a__: int = False for key, mapped_key in MAPPING.items(): if "*" in key: a__ , a__: str = key.split('.*.' ) if prefix in name and suffix in name: a__: List[str] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('embed' ) and name.endswith('embed_avg' ): continue a__: List[str] = True if "*" in mapped_key: a__: List[str] = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: str = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: int = 'weight_g' elif "weight_v" in name: a__: Dict = 'weight_v' elif "weight_ih_l0" in name: a__: int = 'weight_ih_l0' elif "weight_hh_l0" in name: a__: Union[str, Any] = 'weight_hh_l0' elif "bias_ih_l0" in name: a__: Optional[Any] = 'bias_ih_l0' elif "bias_hh_l0" in name: a__: Optional[int] = 'bias_hh_l0' elif "weight_ih_l1" in name: a__: Dict = 'weight_ih_l1' elif "weight_hh_l1" in name: a__: Optional[Any] = 'weight_hh_l1' elif "bias_ih_l1" in name: a__: List[str] = 'bias_ih_l1' elif "bias_hh_l1" in name: a__: Optional[Any] = 'bias_hh_l1' elif "bias" in name: a__: List[str] = 'bias' elif "weight" in name: a__: Any = 'weight' elif "running_mean" in name: a__: Dict = 'running_mean' elif "running_var" in name: a__: Dict = 'running_var' elif "num_batches_tracked" in name: a__: Dict = 'num_batches_tracked' else: a__: List[str] = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->int: if config_path is not None: a__: Dict = EncodecConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: a__: Tuple = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": a__: Any = [8, 5, 4, 4] a__: List[str] = [2.2] a__: List[Any] = 64 a__: Dict = 32000 a__: Union[str, Any] = 2048 a__: Union[str, Any] = False a__: Any = False a__: Optional[Any] = False elif model_name == "encodec_48khz": a__: Optional[int] = [8, 5, 4, 2] a__: Union[str, Any] = [3.0, 6.0, 12.0, 24.0] a__: List[str] = 48000 a__: Tuple = 2 a__: Optional[Any] = False a__: Optional[int] = 'time_group_norm' a__: Union[str, Any] = True a__: Dict = 1.0 a__: str = 0.01 else: raise ValueError(F'Unknown model name: {model_name}' ) a__: Optional[int] = EncodecModel(_SCREAMING_SNAKE_CASE ) a__: List[str] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: int = torch.load(_SCREAMING_SNAKE_CASE ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights a__: str = original_checkpoint['best_state'] recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if repo_id: print('Pushing to the hub...' ) feature_extractor.push_to_hub(_SCREAMING_SNAKE_CASE ) model.push_to_hub(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
290
1
"""simple docstring""" import string import numpy def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: return b if a == 0 else greatest_common_divisor(b % a , _SCREAMING_SNAKE_CASE ) class __snake_case : a__ = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) a__ = numpy.vectorize(lambda __lowerCAmelCase : x % 36 ) a__ = numpy.vectorize(__lowerCAmelCase ) def __init__( self , lowercase) -> None: '''simple docstring''' a__: List[str] = self.modulus(lowercase) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key a__: Dict = encrypt_key.shape[0] def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' return self.key_string.index(lowercase) def lowerCamelCase_ ( self , lowercase) -> str: '''simple docstring''' return self.key_string[round(lowercase)] def lowerCamelCase_ ( self) -> None: '''simple docstring''' a__: int = round(numpy.linalg.det(self.encrypt_key)) if det < 0: a__: Optional[Any] = det % len(self.key_string) a__: Union[str, Any] = len(self.key_string) if greatest_common_divisor(lowercase , len(self.key_string)) != 1: a__: Optional[int] = ( f'determinant modular {req_l} of encryption key({det}) ' f'is not co prime w.r.t {req_l}.\nTry another key.' ) raise ValueError(lowercase) def lowerCamelCase_ ( self , lowercase) -> str: '''simple docstring''' a__: str = [char for char in text.upper() if char in self.key_string] a__: Tuple = chars[-1] while len(lowercase) % self.break_key != 0: chars.append(lowercase) return "".join(lowercase) def lowerCamelCase_ ( self , lowercase) -> str: '''simple docstring''' a__: Any = self.process_text(text.upper()) a__: Optional[int] = '' for i in range(0 , len(lowercase) - self.break_key + 1 , self.break_key): a__: Union[str, Any] = text[i : i + self.break_key] a__: Tuple = [self.replace_letters(lowercase) for char in batch] a__: int = numpy.array([vec]).T a__: int = self.modulus(self.encrypt_key.dot(lowercase)).T.tolist()[ 0 ] a__: Any = ''.join( self.replace_digits(lowercase) for num in batch_encrypted) encrypted += encrypted_batch return encrypted def lowerCamelCase_ ( self) -> numpy.ndarray: '''simple docstring''' a__: int = round(numpy.linalg.det(self.encrypt_key)) if det < 0: a__: Dict = det % len(self.key_string) a__: str = None for i in range(len(self.key_string)): if (det * i) % len(self.key_string) == 1: a__: Any = i break a__: int = ( det_inv * numpy.linalg.det(self.encrypt_key) * numpy.linalg.inv(self.encrypt_key) ) return self.to_int(self.modulus(lowercase)) def lowerCamelCase_ ( self , lowercase) -> str: '''simple docstring''' a__: Tuple = self.make_decrypt_key() a__: Optional[int] = self.process_text(text.upper()) a__: str = '' for i in range(0 , len(lowercase) - self.break_key + 1 , self.break_key): a__: Optional[int] = text[i : i + self.break_key] a__: Optional[int] = [self.replace_letters(lowercase) for char in batch] a__: Any = numpy.array([vec]).T a__: List[Any] = self.modulus(decrypt_key.dot(lowercase)).T.tolist()[0] a__: Tuple = ''.join( self.replace_digits(lowercase) for num in batch_decrypted) decrypted += decrypted_batch return decrypted def __a ( ) ->None: a__: Any = int(input('Enter the order of the encryption key: ' ) ) a__: List[Any] = [] print('Enter each row of the encryption key with space separated integers' ) for _ in range(_SCREAMING_SNAKE_CASE ): a__: List[Any] = [int(_SCREAMING_SNAKE_CASE ) for x in input().split()] hill_matrix.append(_SCREAMING_SNAKE_CASE ) a__: Tuple = HillCipher(numpy.array(_SCREAMING_SNAKE_CASE ) ) print('Would you like to encrypt or decrypt some text? (1 or 2)' ) a__: Any = input('\n1. Encrypt\n2. Decrypt\n' ) if option == "1": a__: Optional[int] = input('What text would you like to encrypt?: ' ) print('Your encrypted text is:' ) print(hc.encrypt(_SCREAMING_SNAKE_CASE ) ) elif option == "2": a__: Any = input('What text would you like to decrypt?: ' ) print('Your decrypted text is:' ) print(hc.decrypt(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: print('moving disk from' , _SCREAMING_SNAKE_CASE , 'to' , _SCREAMING_SNAKE_CASE ) def __a ( ) ->List[str]: a__: Dict = int(input('Height of hanoi: ' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
290
1
"""simple docstring""" from transformers import DistilBertTokenizer, DistilBertTokenizerFast from transformers.testing_utils import require_tokenizers, slow from ..bert.test_tokenization_bert import BertTokenizationTest @require_tokenizers class __snake_case ( __lowerCAmelCase ): a__ = DistilBertTokenizer a__ = DistilBertTokenizerFast a__ = True @slow def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = DistilBertTokenizer.from_pretrained('distilbert-base-uncased') a__: Tuple = tokenizer.encode('sequence builders' , add_special_tokens=lowercase) a__: str = tokenizer.encode('multi-sequence build' , add_special_tokens=lowercase) a__: Dict = tokenizer.build_inputs_with_special_tokens(lowercase) a__: Dict = tokenizer.build_inputs_with_special_tokens(lowercase , lowercase) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ]
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = False ) ->str: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = F'Expected string as input, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[str] = F'Expected boolean as use_pascal parameter, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) a__: int = input_str.split('_' ) a__: List[str] = 0 if use_pascal else 1 a__: List[str] = words[start_index:] a__: List[str] = [word[0].upper() + word[1:] for word in words_to_capitalize] a__: List[str] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ = {'tokenization_bertweet': ['BertweetTokenizer']} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" class __snake_case : def __init__( self , lowercase , lowercase=None , lowercase=None) -> List[str]: '''simple docstring''' a__: Dict = data a__: List[Any] = previous a__: Any = next_node def __str__( self) -> str: '''simple docstring''' return f'{self.data}' def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.data def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.next def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return self.previous class __snake_case : def __init__( self , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = head def __iter__( self) -> List[Any]: '''simple docstring''' return self def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if not self.current: raise StopIteration else: a__: Dict = self.current.get_data() a__: Optional[Any] = self.current.get_next() return value class __snake_case : def __init__( self) -> Dict: '''simple docstring''' a__: List[Any] = None # First node in list a__: Optional[int] = None # Last node in list def __str__( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.head a__: Optional[Any] = [] while current is not None: nodes.append(current.get_data()) a__: str = current.get_next() return " ".join(str(lowercase) for node in nodes) def __contains__( self , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = self.head while current: if current.get_data() == value: return True a__: Dict = current.get_next() return False def __iter__( self) -> int: '''simple docstring''' return LinkedListIterator(self.head) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: a__: Optional[Any] = node a__: Optional[Any] = node else: self.insert_before_node(self.head , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: self.set_head(lowercase) else: self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = Node(lowercase) if self.head is None: self.set_head(lowercase) else: self.set_tail(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Union[str, Any] = node a__: Optional[Any] = node.previous if node.get_previous() is None: a__: Tuple = node_to_insert else: a__: int = node_to_insert a__: Optional[int] = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Optional[int] = node a__: Tuple = node.next if node.get_next() is None: a__: Optional[int] = node_to_insert else: a__: Any = node_to_insert a__: str = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Any = 1 a__: Tuple = Node(lowercase) a__: Tuple = self.head while node: if current_position == position: self.insert_before_node(lowercase , lowercase) return current_position += 1 a__: List[Any] = node.next self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> Node: '''simple docstring''' a__: Tuple = self.head while node: if node.get_data() == item: return node a__: List[str] = node.get_next() raise Exception('Node not found') def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' if (node := self.get_node(lowercase)) is not None: if node == self.head: a__: Any = self.head.get_next() if node == self.tail: a__: List[Any] = self.tail.get_previous() self.remove_node_pointers(lowercase) @staticmethod def lowerCamelCase_ ( lowercase) -> None: '''simple docstring''' if node.get_next(): a__: Any = node.previous if node.get_previous(): a__: List[str] = node.next a__: int = None a__: Union[str, Any] = None def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self.head is None def __a ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ = { 'configuration_mvp': ['MVP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MvpConfig', 'MvpOnnxConfig'], 'tokenization_mvp': ['MvpTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['MvpTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'MVP_PRETRAINED_MODEL_ARCHIVE_LIST', 'MvpForCausalLM', 'MvpForConditionalGeneration', 'MvpForQuestionAnswering', 'MvpForSequenceClassification', 'MvpModel', 'MvpPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE ) ->list: if len(_SCREAMING_SNAKE_CASE ) < 2: return collection def circle_sort_util(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> bool: a__: List[str] = False if low == high: return swapped a__: Tuple = low a__: List[Any] = high while left < right: if collection[left] > collection[right]: a__ , a__: Tuple = ( collection[right], collection[left], ) a__: str = True left += 1 right -= 1 if left == right and collection[left] > collection[right + 1]: a__ , a__: Optional[Any] = ( collection[right + 1], collection[left], ) a__: str = True a__: Any = low + int((high - low) / 2 ) a__: Optional[Any] = circle_sort_util(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Dict = circle_sort_util(_SCREAMING_SNAKE_CASE , mid + 1 , _SCREAMING_SNAKE_CASE ) return swapped or left_swap or right_swap a__: Tuple = True while is_not_sorted is True: a__: Any = circle_sort_util(_SCREAMING_SNAKE_CASE , 0 , len(_SCREAMING_SNAKE_CASE ) - 1 ) return collection if __name__ == "__main__": lowercase__ = input('Enter numbers separated by a comma:\n').strip() lowercase__ = [int(item) for item in user_input.split(',')] print(circle_sort(unsorted))
290
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowercase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Any: if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): a__: Optional[int] = [image] a__: str = [trans(img.convert('RGB' ) ) for img in image] a__: Any = torch.stack(_SCREAMING_SNAKE_CASE ) return image class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM a__: Dict = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowercase , scheduler=lowercase) def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f'The value of strength should in [0.0, 1.0] but is {strength}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: int = min(int(num_inference_steps * strength) , lowercase) a__: Any = max(num_inference_steps - init_timestep , 0) a__: Union[str, Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> List[Any]: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( f'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase)}') a__: Tuple = image.to(device=lowercase , dtype=lowercase) if isinstance(lowercase , lowercase) and len(lowercase) != batch_size: raise ValueError( f'You have passed a list of generators of length {len(lowercase)}, but requested an effective batch' f' size of {batch_size}. Make sure the batch size matches the length of the generators.') a__: List[str] = init_latents.shape a__: List[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) # get latents print('add noise to latents at timestep' , lowercase) a__: int = self.scheduler.add_noise(lowercase , lowercase , lowercase) a__: Dict = init_latents return latents @torch.no_grad() def __call__( self , lowercase = None , lowercase = 0.8 , lowercase = 1 , lowercase = None , lowercase = 0.0 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase) # 2. Preprocess image a__: Tuple = preprocess(lowercase) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device) a__ , a__: Union[str, Any] = self.get_timesteps(lowercase , lowercase , self.device) a__: Optional[int] = timesteps[:1].repeat(lowercase) # 4. Prepare latent variables a__: Union[str, Any] = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase) a__: Optional[Any] = latents # 5. Denoising loop for t in self.progress_bar(lowercase): # 1. predict noise model_output a__: Dict = self.unet(lowercase , lowercase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample a__: Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1) a__: Optional[int] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": a__: Dict = self.numpy_to_pil(lowercase) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase)
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { 'configuration_lxmert': ['LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LxmertConfig'], 'tokenization_lxmert': ['LxmertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['LxmertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'LxmertEncoder', 'LxmertForPreTraining', 'LxmertForQuestionAnswering', 'LxmertModel', 'LxmertPreTrainedModel', 'LxmertVisualFeatureEncoder', 'LxmertXLayer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLxmertForPreTraining', 'TFLxmertMainLayer', 'TFLxmertModel', 'TFLxmertPreTrainedModel', 'TFLxmertVisualFeatureEncoder', ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: Optional[int] = SamImageProcessor() a__: Tuple = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> List[Any]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[Any] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Optional[int] = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: List[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = self.get_image_processor() a__: List[Any] = SamProcessor(image_processor=lowercase) a__: Optional[int] = self.prepare_image_inputs() a__: Optional[Any] = image_processor(lowercase , return_tensors='np') a__: Tuple = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_torch def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.get_image_processor() a__: List[str] = SamProcessor(image_processor=lowercase) a__: Optional[Any] = [torch.ones((1, 3, 5, 5))] a__: Union[str, Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: int = processor.post_process_masks(lowercase , lowercase , lowercase) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Optional[int] = processor.post_process_masks( lowercase , torch.tensor(lowercase) , torch.tensor(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Dict = [np.ones((1, 3, 5, 5))] a__: Tuple = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = [[1, 0], [0, 1]] with self.assertRaises(lowercase): a__: List[Any] = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) @require_vision @require_tf class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: List[Any] = SamImageProcessor() a__: Optional[int] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> int: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[int] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Dict = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = self.get_image_processor() a__: str = SamProcessor(image_processor=lowercase) a__: int = self.prepare_image_inputs() a__: int = image_processor(lowercase , return_tensors='np') a__: Dict = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_tf def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Any = SamProcessor(image_processor=lowercase) a__: str = [tf.ones((1, 3, 5, 5))] a__: List[Any] = [[17_64, 26_46]] a__: List[Any] = [[6_83, 10_24]] a__: List[Any] = processor.post_process_masks(lowercase , lowercase , lowercase , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = processor.post_process_masks( lowercase , tf.convert_to_tensor(lowercase) , tf.convert_to_tensor(lowercase) , return_tensors='tf' , ) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Optional[Any] = [np.ones((1, 3, 5, 5))] a__: int = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: List[str] = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError): a__: Any = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = tempfile.mkdtemp() a__: int = SamImageProcessor() a__: Union[str, Any] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Any = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.get_image_processor() a__: int = SamProcessor(image_processor=lowercase) a__: int = np.random.randint(0 , 2 , size=(1, 3, 5, 5)).astype(np.floataa) a__: Dict = [tf.convert_to_tensor(lowercase)] a__: Union[str, Any] = [torch.tensor(lowercase)] a__: List[Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: Tuple = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='tf') a__: str = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='pt') self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy())) @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Dict = SamProcessor(image_processor=lowercase) a__: Any = self.prepare_image_inputs() a__: List[Any] = image_processor(lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Tuple = processor(images=lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Any = image_processor(lowercase , return_tensors='tf')['pixel_values'].numpy() a__: Any = processor(images=lowercase , return_tensors='tf')['pixel_values'].numpy() self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase))
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE ) ->Any: a__: List[Any] = [0] * len(_SCREAMING_SNAKE_CASE ) a__: int = [] a__: Optional[int] = [] a__: Optional[Any] = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(_SCREAMING_SNAKE_CASE ) ): if indegree[i] == 0: queue.append(_SCREAMING_SNAKE_CASE ) while queue: a__: int = queue.pop(0 ) cnt += 1 topo.append(_SCREAMING_SNAKE_CASE ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(_SCREAMING_SNAKE_CASE ) if cnt != len(_SCREAMING_SNAKE_CASE ): print('Cycle exists' ) else: print(_SCREAMING_SNAKE_CASE ) # Adjacency List of Graph lowercase__ = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
290
"""simple docstring""" from math import pow, sqrt def __a ( *_SCREAMING_SNAKE_CASE ) ->bool: a__: Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) > 0 and all(value > 0.0 for value in values ) return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'facebook/deit-base-distilled-patch16-224': ( 'https://huggingface.co/facebook/deit-base-patch16-224/resolve/main/config.json' ), # See all DeiT models at https://huggingface.co/models?filter=deit } class __snake_case ( __lowerCAmelCase ): a__ = """deit""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=2_24 , lowercase=16 , lowercase=3 , lowercase=True , lowercase=16 , **lowercase , ) -> Optional[int]: '''simple docstring''' super().__init__(**lowercase) a__: Optional[Any] = hidden_size a__: Union[str, Any] = num_hidden_layers a__: Dict = num_attention_heads a__: int = intermediate_size a__: Union[str, Any] = hidden_act a__: Tuple = hidden_dropout_prob a__: Tuple = attention_probs_dropout_prob a__: Any = initializer_range a__: Optional[Any] = layer_norm_eps a__: str = image_size a__: List[str] = patch_size a__: Tuple = num_channels a__: Optional[int] = qkv_bias a__: Optional[Any] = encoder_stride class __snake_case ( __lowerCAmelCase ): a__ = version.parse("""1.11""" ) @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ]) @property def lowerCamelCase_ ( self) -> float: '''simple docstring''' return 1e-4
290
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """roberta-prelayernorm""" def __init__( self , lowercase=5_02_65 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=True , lowercase=None , **lowercase , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase) a__: Union[str, Any] = vocab_size a__: str = hidden_size a__: Tuple = num_hidden_layers a__: List[str] = num_attention_heads a__: Dict = hidden_act a__: int = intermediate_size a__: Tuple = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: Tuple = max_position_embeddings a__: Tuple = type_vocab_size a__: Optional[Any] = initializer_range a__: Tuple = layer_norm_eps a__: Optional[int] = position_embedding_type a__: Any = use_cache a__: Dict = classifier_dropout class __snake_case ( __lowerCAmelCase ): @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": a__: str = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a__: Union[str, Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
290
1
"""simple docstring""" import numpy as np def __a ( _SCREAMING_SNAKE_CASE ) ->np.array: return 1 / (1 + np.exp(-vector )) def __a ( _SCREAMING_SNAKE_CASE ) ->np.array: return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """audio-spectrogram-transformer""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=16 , lowercase=True , lowercase=10 , lowercase=10 , lowercase=10_24 , lowercase=1_28 , **lowercase , ) -> str: '''simple docstring''' super().__init__(**lowercase) a__: Any = hidden_size a__: int = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Any = intermediate_size a__: Union[str, Any] = hidden_act a__: int = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: str = initializer_range a__: Tuple = layer_norm_eps a__: Any = patch_size a__: int = qkv_bias a__: Optional[Any] = frequency_stride a__: int = time_stride a__: List[str] = max_length a__: Tuple = num_mel_bins
290
1
"""simple docstring""" from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
290
"""simple docstring""" import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowercase__ = get_tests_dir('fixtures/test_sentencepiece.model') lowercase__ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') lowercase__ = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = CamembertTokenizer a__ = CamembertTokenizerFast a__ = True a__ = True def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing a__: Tuple = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Optional[Any] = '<pad>' a__: List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase) , lowercase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase) , lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: str = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>NOTUSED') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(lowercase) , 10_04) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_05) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[Any] = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) a__: List[Any] = CamembertTokenizerFast.from_pretrained(self.tmpdirname) a__: Dict = 'I was born in 92000, and this is falsé.' a__: Optional[int] = tokenizer.encode(lowercase) a__: Any = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) a__: Optional[Any] = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) a__: Tuple = tokenizer.convert_ids_to_tokens(lowercase) a__: Tuple = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' if not self.test_rust_tokenizer: return a__: Dict = self.get_tokenizer() a__: str = self.get_rust_tokenizer() a__: int = 'I was born in 92000, and this is falsé.' a__: Optional[Any] = tokenizer.tokenize(lowercase) a__: List[Any] = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) a__: str = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) a__: Tuple = self.get_rust_tokenizer() a__: Union[str, Any] = tokenizer.encode(lowercase) a__: List[Any] = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Union[str, Any] = {'input_ids': [[5, 54, 71_96, 2_97, 30, 23, 7_76, 18, 11, 32_15, 37_05, 82_52, 22, 31_64, 11_81, 21_16, 29, 16, 8_13, 25, 7_91, 33_14, 20, 34_46, 38, 2_75_75, 1_20, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 4_68, 17, 11, 90_88, 20, 15_17, 8, 2_28_04, 1_88_18, 10, 38, 6_29, 6_07, 6_07, 1_42, 19, 71_96, 8_67, 56, 1_03_26, 24, 22_67, 20, 4_16, 50_72, 1_56_12, 2_33, 7_34, 7, 23_99, 27, 16, 30_15, 16_49, 7, 24, 20, 43_38, 23_99, 27, 13, 34_00, 14, 13, 61_89, 8, 9_30, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. a__: int = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=lowercase , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=lowercase , )
290
1
"""simple docstring""" from __future__ import annotations from cmath import sqrt def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->tuple[complex, complex]: if a == 0: raise ValueError('Coefficient \'a\' must not be zero.' ) a__: List[str] = b * b - 4 * a * c a__: Dict = (-b + sqrt(_SCREAMING_SNAKE_CASE )) / (2 * a) a__: str = (-b - sqrt(_SCREAMING_SNAKE_CASE )) / (2 * a) return ( root_a.real if not root_a.imag else root_a, root_a.real if not root_a.imag else root_a, ) def __a ( ) ->int: a__ , a__: Optional[Any] = quadratic_roots(a=5 , b=6 , c=1 ) print(F'The solutions are: {solutiona} and {solutiona}' ) if __name__ == "__main__": main()
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE = 1000000 ) ->int: a__: int = limit + 1 a__: Optional[int] = [0] * limit for first_term in range(1 , _SCREAMING_SNAKE_CASE ): for n in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a a__: Any = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f"{solution() = }")
290
1
"""simple docstring""" from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def __a ( _SCREAMING_SNAKE_CASE ) ->str: if not is_accelerate_available(): return method a__: List[str] = version.parse(accelerate.__version__ ).base_version if version.parse(_SCREAMING_SNAKE_CASE ) < version.parse('0.17.0' ): return method def wrapper(self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): if hasattr(self , '_hf_hook' ) and hasattr(self._hf_hook , 'pre_forward' ): self._hf_hook.pre_forward(self ) return method(self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) return wrapper
290
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union lowercase__ = TypeVar('T') lowercase__ = Union[List[T], Tuple[T, ...]] lowercase__ = Union[T, List[T], Dict[str, T]] lowercase__ = Union[str, bytes, os.PathLike]
290
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor lowercase__ = logging.get_logger(__name__) class __snake_case ( __lowerCAmelCase ): def __init__( self , *lowercase , **lowercase) -> None: '''simple docstring''' warnings.warn( 'The class YolosFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use YolosImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase)
290
"""simple docstring""" from math import pi, sqrt, tan def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) a__: List[Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(_SCREAMING_SNAKE_CASE , 2 ) * torus_radius * tube_radius def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) a__: int = (sidea + sidea + sidea) / 2 a__: Tuple = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(f"Rectangle: {area_rectangle(10, 20) = }") print(f"Square: {area_square(10) = }") print(f"Triangle: {area_triangle(10, 10) = }") print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(f"Parallelogram: {area_parallelogram(10, 20) = }") print(f"Rhombus: {area_rhombus(10, 20) = }") print(f"Trapezium: {area_trapezium(10, 20, 30) = }") print(f"Circle: {area_circle(20) = }") print(f"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(f"Cube: {surface_area_cube(20) = }") print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(f"Sphere: {surface_area_sphere(20) = }") print(f"Hemisphere: {surface_area_hemisphere(20) = }") print(f"Cone: {surface_area_cone(10, 20) = }") print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(f"Cylinder: {surface_area_cylinder(10, 20) = }") print(f"Torus: {surface_area_torus(20, 10) = }") print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(f"Square: {area_reg_polygon(4, 10) = }") print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE ) ->int: if a < 0: raise ValueError('Input value must be a positive integer' ) elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): raise TypeError('Input value must be a \'int\' type' ) return bin(_SCREAMING_SNAKE_CASE ).count('1' ) if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase__ = random.Random() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: if rng is None: a__: Any = global_rng a__: int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=4_00 , lowercase=20_00 , lowercase=1 , lowercase=0.0 , lowercase=1_60_00 , lowercase=True , lowercase=True , ) -> Union[str, Any]: '''simple docstring''' a__: Tuple = parent a__: Optional[int] = batch_size a__: Optional[Any] = min_seq_length a__: Optional[int] = max_seq_length a__: Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a__: Dict = feature_size a__: Any = padding_value a__: Optional[Any] = sampling_rate a__: Optional[Any] = return_attention_mask a__: str = do_normalize def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def lowerCamelCase_ ( self , lowercase=False , lowercase=False) -> Tuple: '''simple docstring''' def _flatten(lowercase): return list(itertools.chain(*lowercase)) if equal_length: a__: Dict = floats_list((self.batch_size, self.max_seq_length)) else: # make sure that inputs increase in size a__: List[Any] = [ _flatten(floats_list((x, self.feature_size))) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: a__: str = [np.asarray(lowercase) for x in speech_inputs] return speech_inputs class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = WavaVecaFeatureExtractor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = WavaVecaFeatureExtractionTester(self) def lowerCamelCase_ ( self , lowercase) -> List[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(lowercase , axis=0) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0) - 1) < 1e-3)) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 a__: Optional[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: List[str] = [np.asarray(lowercase) for speech_input in speech_inputs] # Test not batched input a__: Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='np').input_values a__: Dict = feat_extract(np_speech_inputs[0] , return_tensors='np').input_values self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test batched a__: Dict = feat_extract(lowercase , return_tensors='np').input_values a__: int = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test 2-D numpy arrays are batched. a__: int = [floats_list((1, x))[0] for x in (8_00, 8_00, 8_00)] a__: Union[str, Any] = np.asarray(lowercase) a__: int = feat_extract(lowercase , return_tensors='np').input_values a__: Any = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: List[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Optional[int] = ['longest', 'max_length', 'do_not_pad'] a__: List[Any] = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: Dict = feat_extract(lowercase , padding=lowercase , max_length=lowercase , return_tensors='np') a__: Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self.assertTrue(input_values[0][8_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self.assertTrue(input_values[0][10_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Optional[int] = range(8_00 , 14_00 , 2_00) a__: List[str] = [floats_list((1, x))[0] for x in lengths] a__: Tuple = ['longest', 'max_length', 'do_not_pad'] a__: Dict = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: int = feat_extract(lowercase , max_length=lowercase , padding=lowercase) a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Any = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Dict = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='max_length' , return_tensors='np') a__: int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1]) self._check_zero_mean_unit_variance(input_values[2]) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: int = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: str = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='longest' , return_tensors='np') a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00)) a__: Dict = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Tuple = feat_extract( lowercase , truncation=lowercase , max_length=20_00 , padding='longest' , return_tensors='np') a__: str = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00)) @require_torch def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' import torch a__: Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Tuple = np.random.rand(1_00).astype(np.floataa) a__: Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a__: Any = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np') self.assertTrue(np_processed.input_values.dtype == np.floataa) a__: Optional[Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt') self.assertTrue(pt_processed.input_values.dtype == torch.floataa) @slow @require_torch def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: a__: str = WavaVecaConfig.from_pretrained(lowercase) a__: str = WavaVecaFeatureExtractor.from_pretrained(lowercase) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer')
290
1
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) def __a ( _SCREAMING_SNAKE_CASE ) ->int: a__: Dict = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): a__: int = key.replace('module.encoder' , 'glpn.encoder' ) if key.startswith('module.decoder' ): a__: Any = key.replace('module.decoder' , 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 a__: List[Any] = key[key.find('patch_embed' ) + len('patch_embed' )] a__: Tuple = key.replace(F'patch_embed{idx}' , F'patch_embeddings.{int(_SCREAMING_SNAKE_CASE )-1}' ) if "norm" in key: a__: Union[str, Any] = key.replace('norm' , 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 a__: Optional[Any] = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] a__: Optional[Any] = key.replace(F'layer_norm{idx}' , F'layer_norm.{int(_SCREAMING_SNAKE_CASE )-1}' ) if "layer_norm1" in key: a__: Any = key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: a__: Any = key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 a__: str = key[key.find('block' ) + len('block' )] a__: str = key.replace(F'block{idx}' , F'block.{int(_SCREAMING_SNAKE_CASE )-1}' ) if "attn.q" in key: a__: Optional[int] = key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: a__: Optional[int] = key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: a__: Tuple = key.replace('attn' , 'attention.self' ) if "fc1" in key: a__: int = key.replace('fc1' , 'dense1' ) if "fc2" in key: a__: List[Any] = key.replace('fc2' , 'dense2' ) if "linear_pred" in key: a__: int = key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: a__: Optional[int] = key.replace('linear_fuse.conv' , 'linear_fuse' ) a__: int = key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 a__: Dict = key[key.find('linear_c' ) + len('linear_c' )] a__: str = key.replace(F'linear_c{idx}' , F'linear_c.{int(_SCREAMING_SNAKE_CASE )-1}' ) if "bot_conv" in key: a__: Dict = key.replace('bot_conv' , '0.convolution' ) if "skip_conv1" in key: a__: Any = key.replace('skip_conv1' , '1.convolution' ) if "skip_conv2" in key: a__: Optional[Any] = key.replace('skip_conv2' , '2.convolution' ) if "fusion1" in key: a__: Tuple = key.replace('fusion1' , '1.fusion' ) if "fusion2" in key: a__: Optional[Any] = key.replace('fusion2' , '2.fusion' ) if "fusion3" in key: a__: Optional[int] = key.replace('fusion3' , '3.fusion' ) if "fusion" in key and "conv" in key: a__: Union[str, Any] = key.replace('conv' , 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): a__: Optional[Any] = key.replace('module.last_layer_depth' , 'head.head' ) a__: Optional[Any] = value return new_state_dict def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) a__: str = state_dict.pop(F'glpn.encoder.block.{i}.{j}.attention.self.kv.weight' ) a__: Tuple = state_dict.pop(F'glpn.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict a__: Tuple = kv_weight[ : config.hidden_sizes[i], : ] a__: List[Any] = kv_bias[: config.hidden_sizes[i]] a__: Dict = kv_weight[ config.hidden_sizes[i] :, : ] a__: str = kv_bias[config.hidden_sizes[i] :] def __a ( ) ->str: a__: Optional[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' a__: Union[str, Any] = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ) return image @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: Optional[int] = GLPNConfig(hidden_sizes=[64, 128, 320, 512] , decoder_hidden_size=64 , depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) a__: Optional[Any] = GLPNImageProcessor() # prepare image a__: Dict = prepare_img() a__: Union[str, Any] = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict a__: Optional[Any] = torch.load(_SCREAMING_SNAKE_CASE , map_location=torch.device('cpu' ) ) # rename keys a__: List[str] = rename_keys(_SCREAMING_SNAKE_CASE ) # key and value matrices need special treatment read_in_k_v(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # create HuggingFace model and load state dict a__: Optional[int] = GLPNForDepthEstimation(_SCREAMING_SNAKE_CASE ) model.load_state_dict(_SCREAMING_SNAKE_CASE ) model.eval() # forward pass a__: int = model(_SCREAMING_SNAKE_CASE ) a__: Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: a__: Dict = torch.tensor( [[4.4_147, 4.0_873, 4.0_673], [3.7_890, 3.2_881, 3.1_525], [3.7_674, 3.5_423, 3.4_913]] ) elif "kitti" in model_name: a__: List[Any] = torch.tensor( [[3.4_291, 2.7_865, 2.5_151], [3.2_841, 2.7_021, 2.3_502], [3.1_147, 2.4_625, 2.2_481]] ) else: raise ValueError(F'Unknown model name: {model_name}' ) a__: Optional[int] = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , organization='nielsr' , commit_message='Add model' , use_temp_dir=_SCREAMING_SNAKE_CASE , ) image_processor.push_to_hub( repo_path_or_name=Path(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , organization='nielsr' , commit_message='Add image processor' , use_temp_dir=_SCREAMING_SNAKE_CASE , ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) parser.add_argument( '--model_name', default='glpn-kitti', type=str, help='Name of the model in case you\'re pushing to the hub.', ) lowercase__ = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class __snake_case ( __lowerCAmelCase ): a__ = """decision_transformer""" a__ = ["""past_key_values"""] a__ = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=17 , lowercase=4 , lowercase=1_28 , lowercase=40_96 , lowercase=True , lowercase=1 , lowercase=10_24 , lowercase=3 , lowercase=1 , lowercase=None , lowercase="relu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=1e-5 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=5_02_56 , lowercase=5_02_56 , lowercase=False , lowercase=False , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = state_dim a__: int = act_dim a__: List[Any] = hidden_size a__: List[str] = max_ep_len a__: List[Any] = action_tanh a__: Optional[Any] = vocab_size a__: Tuple = n_positions a__: Dict = n_layer a__: Optional[int] = n_head a__: Optional[int] = n_inner a__: Any = activation_function a__: Union[str, Any] = resid_pdrop a__: Any = embd_pdrop a__: Any = attn_pdrop a__: List[Any] = layer_norm_epsilon a__: Optional[Any] = initializer_range a__: Any = scale_attn_weights a__: Dict = use_cache a__: Optional[int] = scale_attn_by_inverse_layer_idx a__: List[str] = reorder_and_upcast_attn a__: Any = bos_token_id a__: int = eos_token_id super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase)
290
1
"""simple docstring""" import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = ConsistencyModelPipeline a__ = UNCONDITIONAL_IMAGE_GENERATION_PARAMS a__ = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt a__ = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Optional[int] = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet' , ) return unet @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Tuple = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test' , subfolder='test_unet_class_cond' , ) return unet def lowerCamelCase_ ( self , lowercase=False) -> Optional[Any]: '''simple docstring''' if class_cond: a__: Optional[Any] = self.dummy_cond_unet else: a__: int = self.dummy_uncond_unet # Default to CM multistep sampler a__: Optional[int] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a__: List[str] = { 'unet': unet, 'scheduler': scheduler, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> Tuple: '''simple docstring''' if str(lowercase).startswith('mps'): a__: Optional[Any] = torch.manual_seed(lowercase) else: a__: List[Any] = torch.Generator(device=lowercase).manual_seed(lowercase) a__: List[str] = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: List[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator a__: List[str] = self.get_dummy_components() a__: Any = ConsistencyModelPipeline(**lowercase) a__: Optional[Any] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: str = self.get_dummy_inputs(lowercase) a__: Any = pipe(**lowercase).images assert image.shape == (1, 32, 32, 3) a__: List[Any] = image[0, -3:, -3:, -1] a__: Union[str, Any] = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: str = 'cpu' # ensure determinism for the device-dependent torch.Generator a__: Optional[int] = self.get_dummy_components(class_cond=lowercase) a__: Union[str, Any] = ConsistencyModelPipeline(**lowercase) a__: Optional[Any] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: str = self.get_dummy_inputs(lowercase) a__: List[str] = 0 a__: Optional[int] = pipe(**lowercase).images assert image.shape == (1, 32, 32, 3) a__: Optional[int] = image[0, -3:, -3:, -1] a__: List[Any] = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator a__: Union[str, Any] = self.get_dummy_components() a__: Dict = ConsistencyModelPipeline(**lowercase) a__: Dict = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Dict = self.get_dummy_inputs(lowercase) a__: List[Any] = 1 a__: Tuple = None a__: Tuple = pipe(**lowercase).images assert image.shape == (1, 32, 32, 3) a__: List[Any] = image[0, -3:, -3:, -1] a__: Dict = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = 'cpu' # ensure determinism for the device-dependent torch.Generator a__: Union[str, Any] = self.get_dummy_components(class_cond=lowercase) a__: int = ConsistencyModelPipeline(**lowercase) a__: str = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: int = self.get_dummy_inputs(lowercase) a__: Tuple = 1 a__: List[str] = None a__: Tuple = 0 a__: Dict = pipe(**lowercase).images assert image.shape == (1, 32, 32, 3) a__: List[Any] = image[0, -3:, -3:, -1] a__: Dict = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self , lowercase=0 , lowercase=False , lowercase="cpu" , lowercase=torch.floataa , lowercase=(1, 3, 64, 64)) -> Dict: '''simple docstring''' a__: Dict = torch.manual_seed(lowercase) a__: Union[str, Any] = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: a__: List[Any] = self.get_fixed_latents(seed=lowercase , device=lowercase , dtype=lowercase , shape=lowercase) a__: Any = latents return inputs def lowerCamelCase_ ( self , lowercase=0 , lowercase="cpu" , lowercase=torch.floataa , lowercase=(1, 3, 64, 64)) -> Union[str, Any]: '''simple docstring''' if type(lowercase) == str: a__: List[str] = torch.device(lowercase) a__: Dict = torch.Generator(device=lowercase).manual_seed(lowercase) a__: List[str] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) return latents def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Union[str, Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2') a__: Union[str, Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a__: Optional[Any] = ConsistencyModelPipeline(unet=lowercase , scheduler=lowercase) pipe.to(torch_device=lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: List[str] = self.get_inputs() a__: List[Any] = pipe(**lowercase).images assert image.shape == (1, 64, 64, 3) a__: str = image[0, -3:, -3:, -1] a__: List[str] = np.array([0.0888, 0.0881, 0.0666, 0.0479, 0.0292, 0.0195, 0.0201, 0.0163, 0.0254]) assert np.abs(image_slice.flatten() - expected_slice).max() < 2e-2 def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2') a__: Any = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a__: List[Any] = ConsistencyModelPipeline(unet=lowercase , scheduler=lowercase) pipe.to(torch_device=lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[Any] = self.get_inputs() a__: int = 1 a__: str = None a__: Union[str, Any] = pipe(**lowercase).images assert image.shape == (1, 64, 64, 3) a__: Dict = image[0, -3:, -3:, -1] a__: Optional[int] = np.array([0.0340, 0.0152, 0.0063, 0.0267, 0.0221, 0.0107, 0.0416, 0.0186, 0.0217]) assert np.abs(image_slice.flatten() - expected_slice).max() < 2e-2 @require_torch_a def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2') a__: Optional[int] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a__: List[str] = ConsistencyModelPipeline(unet=lowercase , scheduler=lowercase) pipe.to(torch_device=lowercase , torch_dtype=torch.floataa) pipe.set_progress_bar_config(disable=lowercase) a__: int = self.get_inputs(get_fixed_latents=lowercase , device=lowercase) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=lowercase , enable_math=lowercase , enable_mem_efficient=lowercase): a__: Tuple = pipe(**lowercase).images assert image.shape == (1, 64, 64, 3) a__: Union[str, Any] = image[0, -3:, -3:, -1] a__: Dict = np.array([0.1875, 0.1428, 0.1289, 0.2151, 0.2092, 0.1477, 0.1877, 0.1641, 0.1353]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3 @require_torch_a def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: str = UNetaDModel.from_pretrained('diffusers/consistency_models' , subfolder='diffusers_cd_imagenet64_l2') a__: Union[str, Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) a__: Optional[Any] = ConsistencyModelPipeline(unet=lowercase , scheduler=lowercase) pipe.to(torch_device=lowercase , torch_dtype=torch.floataa) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[Any] = self.get_inputs(get_fixed_latents=lowercase , device=lowercase) a__: Union[str, Any] = 1 a__: Tuple = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=lowercase , enable_math=lowercase , enable_mem_efficient=lowercase): a__: Optional[int] = pipe(**lowercase).images assert image.shape == (1, 64, 64, 3) a__: Any = image[0, -3:, -3:, -1] a__: List[Any] = np.array([0.1663, 0.1948, 0.2275, 0.1680, 0.1204, 0.1245, 0.1858, 0.1338, 0.2095]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-3
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: while a != 0: a__ , a__: List[str] = b % a, a return b def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: if gcd(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) != 1: a__: Dict = F'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Union[str, Any] = 1, 0, a a__ , a__ , a__: Any = 0, 1, m while va != 0: a__: int = ua // va a__ , a__ , a__ , a__ , a__ , a__: Any = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
290
1
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union lowercase__ = TypeVar('T') lowercase__ = Union[List[T], Tuple[T, ...]] lowercase__ = Union[T, List[T], Dict[str, T]] lowercase__ = Union[str, bytes, os.PathLike]
290
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex lowercase__ = logging.getLogger(__name__) class __snake_case : def __init__( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' if not self.initialized: a__: Optional[int] = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Optional[int] = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.retriever.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ , a__: str = self.retriever._main_retrieve(lowercase , lowercase) return doc_ids, retrieved_doc_embeds class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> int: '''simple docstring''' if index is not None and index.is_initialized() and len(lowercase) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ') super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Any = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase) for worker in self.retrieval_workers ]) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' logger.info('initializing retrieval') if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if len(self.retrieval_workers) > 0: # Select a random retrieval actor. a__: int = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] a__ , a__: List[Any] = ray.get(random_worker.retrieve.remote(lowercase , lowercase)) else: a__ , a__: Dict = self._main_retrieve(lowercase , lowercase) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' return super(lowercase , cls).get_tokenizers(lowercase , lowercase , **lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = kwargs.pop('config' , lowercase) or RagConfig.from_pretrained(lowercase , **lowercase) a__: Union[str, Any] = RagTokenizer.from_pretrained(lowercase , config=lowercase) a__: int = rag_tokenizer.question_encoder a__: Any = rag_tokenizer.generator if indexed_dataset is not None: a__: List[Any] = 'custom' a__: Optional[Any] = CustomHFIndex(config.retrieval_vector_size , lowercase) else: a__: Dict = cls._build_index(lowercase) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
290
1
"""simple docstring""" from maths.prime_check import is_prime def __a ( _SCREAMING_SNAKE_CASE ) ->int: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Any = F'Input value of [number={number}] must be an integer' raise TypeError(_SCREAMING_SNAKE_CASE ) if is_prime(_SCREAMING_SNAKE_CASE ) and is_prime(number + 2 ): return number + 2 else: return -1 if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: int = None if token is not None: a__: Tuple = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Optional[Any] = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: str = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) a__: int = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Dict = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Dict: a__: Dict = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Dict = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: List[Any] = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Optional[int] = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: List[Any] = None if token is not None: a__: Optional[int] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = result.headers['Location'] a__: Optional[int] = requests.get(_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: int = os.path.join(_SCREAMING_SNAKE_CASE , F'{artifact_name}.zip' ) with open(_SCREAMING_SNAKE_CASE , 'wb' ) as fp: fp.write(response.content ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: a__: List[Any] = [] a__: Optional[Any] = [] a__: List[Any] = None with zipfile.ZipFile(_SCREAMING_SNAKE_CASE ) as z: for filename in z.namelist(): if not os.path.isdir(_SCREAMING_SNAKE_CASE ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(_SCREAMING_SNAKE_CASE ) as f: for line in f: a__: Optional[int] = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs a__: Union[str, Any] = line[: line.index(': ' )] a__: Union[str, Any] = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed a__: Optional[int] = line[len('FAILED ' ) :] failed_tests.append(_SCREAMING_SNAKE_CASE ) elif filename == "job_name.txt": a__: Union[str, Any] = line if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( F'`errors` and `failed_tests` should have the same number of elements. Got {len(_SCREAMING_SNAKE_CASE )} for `errors` ' F'and {len(_SCREAMING_SNAKE_CASE )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some' ' problem.' ) a__: Tuple = None if job_name and job_links: a__: Dict = job_links.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # A list with elements of the form (line of error, error, failed test) a__: int = [x + [y] + [job_link] for x, y in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )] return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->str: a__: int = [] a__: Optional[int] = [os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for p in os.listdir(_SCREAMING_SNAKE_CASE ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(_SCREAMING_SNAKE_CASE , job_links=_SCREAMING_SNAKE_CASE ) ) return errors def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Any: a__: str = Counter() counter.update([x[1] for x in logs] ) a__: int = counter.most_common() a__: Any = {} for error, count in counts: if error_filter is None or error not in error_filter: a__: List[str] = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} a__: Optional[Any] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: List[str] = test.split('::' )[0] if test.startswith('tests/models/' ): a__: Dict = test.split('/' )[2] else: a__: Any = None return test def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: int = [(x[0], x[1], get_model(x[2] )) for x in logs] a__: List[Any] = [x for x in logs if x[2] is not None] a__: Optional[Any] = {x[2] for x in logs} a__: Dict = {} for test in tests: a__: Union[str, Any] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) a__: Union[str, Any] = counter.most_common() a__: List[str] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} a__: List[Any] = sum(error_counts.values() ) if n_errors > 0: a__: Any = {'count': n_errors, 'errors': error_counts} a__: Optional[int] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Any = '| no. | error | status |' a__: Any = '|-:|:-|:-|' a__: str = [header, sep] for error in reduced_by_error: a__: int = reduced_by_error[error]['count'] a__: Tuple = F'| {count} | {error[:100]} | |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->str: a__: List[str] = '| model | no. of errors | major error | count |' a__: str = '|-:|-:|-:|-:|' a__: int = [header, sep] for model in reduced_by_model: a__: Tuple = reduced_by_model[model]['count'] a__ , a__: Dict = list(reduced_by_model[model]['errors'].items() )[0] a__: Dict = F'| {model} | {count} | {error[:60]} | {_count} |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') lowercase__ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowercase__ = get_job_links(args.workflow_run_id, token=args.token) lowercase__ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowercase__ = k.find(' / ') lowercase__ = k[index + len(' / ') :] lowercase__ = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowercase__ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowercase__ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowercase__ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowercase__ = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowercase__ = reduce_by_error(errors) lowercase__ = reduce_by_model(errors) lowercase__ = make_github_table(reduced_by_error) lowercase__ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: while b: a__ , a__: List[str] = b, a % b return a def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: return a if b == 0 else euclidean_gcd_recursive(_SCREAMING_SNAKE_CASE , a % b ) def __a ( ) ->Union[str, Any]: print(F'euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}' ) print(F'euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}' ) print(F'euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}' ) print(F'euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}' ) print(F'euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}' ) print(F'euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}' ) print(F'euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}' ) print(F'euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}' ) print(F'euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}' ) print(F'euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}' ) if __name__ == "__main__": main()
290
"""simple docstring""" import math def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __a ( _SCREAMING_SNAKE_CASE = 0.1 ) ->int: a__: str = 3 a__: Optional[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_SCREAMING_SNAKE_CASE ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" import collections import importlib.util import os import re from pathlib import Path lowercase__ = 'src/transformers' # Matches is_xxx_available() lowercase__ = re.compile(r'is\_([a-z_]*)_available()') # Catches a one-line _import_struct = {xxx} lowercase__ = re.compile(r'^_import_structure\s+=\s+\{([^\}]+)\}') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowercase__ = re.compile(r'\s+"\S*":\s+\[([^\]]*)\]') # Catches a line if not is_foo_available lowercase__ = re.compile(r'^\s*if\s+not\s+is\_[a-z_]*\_available\(\)') # Catches a line _import_struct["bla"].append("foo") lowercase__ = re.compile(r'^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowercase__ = re.compile(r'^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]') # Catches a line with an object between quotes and a comma: "MyModel", lowercase__ = re.compile('^\s+"([^"]+)",') # Catches a line with objects between brackets only: ["foo", "bar"], lowercase__ = re.compile('^\s+\[([^\]]+)\]') # Catches a line with from foo import bar, bla, boo lowercase__ = re.compile(r'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') # Catches a line with try: lowercase__ = re.compile(r'^\s*try:') # Catches a line with else: lowercase__ = re.compile(r'^\s*else:') def __a ( _SCREAMING_SNAKE_CASE ) ->str: if _re_test_backend.search(_SCREAMING_SNAKE_CASE ) is None: return None a__: Optional[Any] = [b[0] for b in _re_backend.findall(_SCREAMING_SNAKE_CASE )] backends.sort() return "_and_".join(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->Optional[Any]: with open(_SCREAMING_SNAKE_CASE , 'r' , encoding='utf-8' , newline='\n' ) as f: a__: List[Any] = f.readlines() a__: int = 0 while line_index < len(_SCREAMING_SNAKE_CASE ) and not lines[line_index].startswith('_import_structure = {' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(_SCREAMING_SNAKE_CASE ): return None # First grab the objects without a specific backend in _import_structure a__: int = [] while not lines[line_index].startswith('if TYPE_CHECKING' ) and find_backend(lines[line_index] ) is None: a__: Dict = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(_SCREAMING_SNAKE_CASE ): a__: int = _re_one_line_import_struct.search(_SCREAMING_SNAKE_CASE ).groups()[0] a__: Any = re.findall('\[([^\]]+)\]' , _SCREAMING_SNAKE_CASE ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ' )] ) line_index += 1 continue a__: List[Any] = _re_import_struct_key_value.search(_SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: a__: Union[str, Any] = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ' ) if len(_SCREAMING_SNAKE_CASE ) > 0] objects.extend(_SCREAMING_SNAKE_CASE ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) line_index += 1 a__: List[Any] = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING' ): # If the line is an if not is_backend_available, we grab all objects associated. a__: int = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: a__: Dict = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 a__: List[str] = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 4 ): a__: Optional[Any] = lines[line_index] if _re_import_struct_add_one.search(_SCREAMING_SNAKE_CASE ) is not None: objects.append(_re_import_struct_add_one.search(_SCREAMING_SNAKE_CASE ).groups()[0] ) elif _re_import_struct_add_many.search(_SCREAMING_SNAKE_CASE ) is not None: a__: List[str] = _re_import_struct_add_many.search(_SCREAMING_SNAKE_CASE ).groups()[0].split(', ' ) a__: str = [obj[1:-1] for obj in imports if len(_SCREAMING_SNAKE_CASE ) > 0] objects.extend(_SCREAMING_SNAKE_CASE ) elif _re_between_brackets.search(_SCREAMING_SNAKE_CASE ) is not None: a__: Optional[Any] = _re_between_brackets.search(_SCREAMING_SNAKE_CASE ).groups()[0].split(', ' ) a__: Optional[Any] = [obj[1:-1] for obj in imports if len(_SCREAMING_SNAKE_CASE ) > 0] objects.extend(_SCREAMING_SNAKE_CASE ) elif _re_quote_object.search(_SCREAMING_SNAKE_CASE ) is not None: objects.append(_re_quote_object.search(_SCREAMING_SNAKE_CASE ).groups()[0] ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) elif line.startswith(' ' * 12 + '"' ): objects.append(line[13:-3] ) line_index += 1 a__: Tuple = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend a__: Any = [] while ( line_index < len(_SCREAMING_SNAKE_CASE ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('else' ) ): a__: Optional[int] = lines[line_index] a__: Optional[Any] = _re_import.search(_SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 8 ): objects.append(line[8:-2] ) line_index += 1 a__: str = {'none': objects} # Let's continue with backend-specific objects while line_index < len(_SCREAMING_SNAKE_CASE ): # If the line is an if is_backend_available, we grab all objects associated. a__: Tuple = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: a__: Dict = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 a__: Any = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 8 ): a__: Optional[int] = lines[line_index] a__: Optional[int] = _re_import.search(_SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 12 ): objects.append(line[12:-2] ) line_index += 1 a__: Any = objects else: line_index += 1 return import_dict_objects, type_hint_objects def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def find_duplicates(_SCREAMING_SNAKE_CASE ): return [k for k, v in collections.Counter(_SCREAMING_SNAKE_CASE ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] a__: int = [] for key in import_dict_objects.keys(): a__: Tuple = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F'Duplicate _import_structure definitions for: {duplicate_imports}' ) a__: Optional[int] = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F'Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): a__: Optional[int] = 'base imports' if key == 'none' else F'{key} backend' errors.append(F'Differences for {name}:' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F' {a} in TYPE_HINT but not in _import_structure.' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F' {a} in _import_structure but not in TYPE_HINT.' ) return errors def __a ( ) ->Dict: a__: List[str] = [] for root, _, files in os.walk(_SCREAMING_SNAKE_CASE ): if "__init__.py" in files: a__: Optional[Any] = os.path.join(_SCREAMING_SNAKE_CASE , '__init__.py' ) a__: Any = parse_init(_SCREAMING_SNAKE_CASE ) if objects is not None: a__: Any = analyze_results(*_SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: a__: Dict = F'Problem in {fname}, both halves do not define the same objects.\n{errors[0]}' failures.append('\n'.join(_SCREAMING_SNAKE_CASE ) ) if len(_SCREAMING_SNAKE_CASE ) > 0: raise ValueError('\n\n'.join(_SCREAMING_SNAKE_CASE ) ) def __a ( ) ->Tuple: a__: str = [] for path, directories, files in os.walk(_SCREAMING_SNAKE_CASE ): for folder in directories: # Ignore private modules if folder.startswith('_' ): directories.remove(_SCREAMING_SNAKE_CASE ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(_SCREAMING_SNAKE_CASE ) / folder).glob('*.py' ) ) ) == 0: continue a__: Dict = str((Path(_SCREAMING_SNAKE_CASE ) / folder).relative_to(_SCREAMING_SNAKE_CASE ) ) a__: Tuple = short_path.replace(os.path.sep , '.' ) submodules.append(_SCREAMING_SNAKE_CASE ) for fname in files: if fname == "__init__.py": continue a__: Union[str, Any] = str((Path(_SCREAMING_SNAKE_CASE ) / fname).relative_to(_SCREAMING_SNAKE_CASE ) ) a__: List[Any] = short_path.replace('.py' , '' ).replace(os.path.sep , '.' ) if len(submodule.split('.' ) ) == 1: submodules.append(_SCREAMING_SNAKE_CASE ) return submodules lowercase__ = [ 'convert_pytorch_checkpoint_to_tf2', 'modeling_flax_pytorch_utils', ] def __a ( ) ->List[str]: # This is to make sure the transformers module imported is the one in the repo. a__: Tuple = importlib.util.spec_from_file_location( 'transformers' , os.path.join(_SCREAMING_SNAKE_CASE , '__init__.py' ) , submodule_search_locations=[PATH_TO_TRANSFORMERS] , ) a__: Dict = spec.loader.load_module() a__: int = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(_SCREAMING_SNAKE_CASE ) > 0: a__: List[Any] = '\n'.join(F'- {module}' for module in module_not_registered ) raise ValueError( 'The following submodules are not properly registered in the main init of Transformers:\n' F'{list_of_modules}\n' 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.' ) if __name__ == "__main__": check_all_inits() check_submodules()
290
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ = {'processing_wav2vec2_with_lm': ['Wav2Vec2ProcessorWithLM']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
1
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping lowercase__ = tuple[int, int] class __snake_case : def __init__( self , lowercase , lowercase) -> None: '''simple docstring''' a__: set[int] = vertices a__: dict[EdgeT, int] = { (min(lowercase), max(lowercase)): weight for edge, weight in edges.items() } def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' self.vertices.add(edge[0]) self.vertices.add(edge[1]) a__: Union[str, Any] = weight def lowerCamelCase_ ( self) -> Graph: '''simple docstring''' a__: Graph = Graph({min(self.vertices)} , {}) a__: EdgeT a__: int a__: EdgeT a__: int while len(subgraph.vertices) < len(self.vertices): a__: Union[str, Any] = max(self.edges.values()) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: a__: List[str] = edge a__: Tuple = weight subgraph.add_edge(lowercase , lowercase) return subgraph def __a ( _SCREAMING_SNAKE_CASE = "p107_network.txt" ) ->int: a__: str = os.path.abspath(os.path.dirname(_SCREAMING_SNAKE_CASE ) ) a__: str = os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: dict[EdgeT, int] = {} a__: list[str] a__: int a__: int with open(_SCREAMING_SNAKE_CASE ) as f: a__: Any = f.read().strip().split('\n' ) a__: str = [line.split(',' ) for line in data] for edgea in range(1 , len(_SCREAMING_SNAKE_CASE ) ): for edgea in range(_SCREAMING_SNAKE_CASE ): if adjaceny_matrix[edgea][edgea] != "-": a__: List[str] = int(adjaceny_matrix[edgea][edgea] ) a__: Graph = Graph(set(range(len(_SCREAMING_SNAKE_CASE ) ) ) , _SCREAMING_SNAKE_CASE ) a__: Graph = graph.prims_algorithm() a__: int = sum(graph.edges.values() ) a__: int = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(f"{solution() = }")
290
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = KandinskyInpaintPipeline a__ = ["""prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] a__ = [ """prompt""", """negative_prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] a__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """negative_prompt""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] a__ = False @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 1_00 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[int] = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base') return tokenizer @property def lowerCamelCase_ ( self) -> Any: '''simple docstring''' torch.manual_seed(0) a__: Dict = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=10_05 , ) a__: Optional[Any] = MultilingualCLIP(lowercase) a__: int = text_encoder.eval() return text_encoder @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' torch.manual_seed(0) a__: Any = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } a__: str = UNetaDConditionModel(**lowercase) return model @property def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' torch.manual_seed(0) a__: Any = VQModel(**self.dummy_movq_kwargs) return model def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = self.dummy_text_encoder a__: int = self.dummy_tokenizer a__: str = self.dummy_unet a__: Any = self.dummy_movq a__: Tuple = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.00085 , beta_end=0.012 , clip_sample=lowercase , set_alpha_to_one=lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=lowercase , ) a__: Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> Any: '''simple docstring''' a__: List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase)).to(lowercase) a__: int = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1)).to(lowercase) # create init_image a__: Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) a__: int = image.cpu().permute(0 , 2 , 3 , 1)[0] a__: Optional[int] = Image.fromarray(np.uinta(lowercase)).convert('RGB').resize((2_56, 2_56)) # create mask a__: Tuple = np.ones((64, 64) , dtype=np.floataa) a__: Optional[Any] = 0 if str(lowercase).startswith('mps'): a__: str = torch.manual_seed(lowercase) else: a__: Dict = torch.Generator(device=lowercase).manual_seed(lowercase) a__: Optional[int] = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[Any] = 'cpu' a__: List[Any] = self.get_dummy_components() a__: Optional[Any] = self.pipeline_class(**lowercase) a__: str = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[int] = pipe(**self.get_dummy_inputs(lowercase)) a__: List[str] = output.images a__: int = pipe( **self.get_dummy_inputs(lowercase) , return_dict=lowercase , )[0] a__: Optional[Any] = image[0, -3:, -3:, -1] a__: List[Any] = image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}') assert image.shape == (1, 64, 64, 3) a__: str = np.array( [0.8326919, 0.73790467, 0.20918581, 0.9309612, 0.5511791, 0.43713328, 0.5513321, 0.49922934, 0.59497786]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy') a__: int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png') a__: Union[str, Any] = np.ones((7_68, 7_68) , dtype=np.floataa) a__: int = 0 a__: Optional[int] = 'a hat' a__: int = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa) pipe_prior.to(lowercase) a__: Any = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa) a__: Optional[Any] = pipeline.to(lowercase) pipeline.set_progress_bar_config(disable=lowercase) a__: Dict = torch.Generator(device='cpu').manual_seed(0) a__ , a__: Optional[Any] = pipe_prior( lowercase , generator=lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() a__: List[str] = pipeline( lowercase , image=lowercase , mask_image=lowercase , image_embeds=lowercase , negative_image_embeds=lowercase , generator=lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) a__: str = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(lowercase , lowercase)
290
1
"""simple docstring""" import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def __a ( *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=2 ) ->Tuple: from .. import __version__ a__: Dict = take_from a__: Tuple = () if not isinstance(args[0] , _SCREAMING_SNAKE_CASE ): a__: int = (args,) for attribute, version_name, message in args: if version.parse(version.parse(_SCREAMING_SNAKE_CASE ).base_version ) >= version.parse(_SCREAMING_SNAKE_CASE ): raise ValueError( F'The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers\'' F' version {__version__} is >= {version_name}' ) a__: Optional[Any] = None if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(_SCREAMING_SNAKE_CASE ),) a__: List[Any] = F'The `{attribute}` argument is deprecated and will be removed in version {version_name}.' elif hasattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): values += (getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ),) a__: Optional[int] = F'The `{attribute}` attribute is deprecated and will be removed in version {version_name}.' elif deprecated_kwargs is None: a__: Dict = F'`{attribute}` is deprecated and will be removed in version {version_name}.' if warning is not None: a__: Union[str, Any] = warning + ' ' if standard_warn else '' warnings.warn(warning + message , _SCREAMING_SNAKE_CASE , stacklevel=_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and len(_SCREAMING_SNAKE_CASE ) > 0: a__: Optional[Any] = inspect.getouterframes(inspect.currentframe() )[1] a__: Dict = call_frame.filename a__: Union[str, Any] = call_frame.lineno a__: int = call_frame.function a__ , a__: Any = next(iter(deprecated_kwargs.items() ) ) raise TypeError(F'{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`' ) if len(_SCREAMING_SNAKE_CASE ) == 0: return elif len(_SCREAMING_SNAKE_CASE ) == 1: return values[0] return values
290
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() lowercase__ = logging.get_logger('transformers.models.encodec') lowercase__ = { 'quantizer.vq.layers.*._codebook.inited': 'quantizer.layers.*.codebook.inited', 'quantizer.vq.layers.*._codebook.cluster_size': 'quantizer.layers.*.codebook.cluster_size', 'quantizer.vq.layers.*._codebook.embed': 'quantizer.layers.*.codebook.embed', 'quantizer.vq.layers.*._codebook.embed_avg': 'quantizer.layers.*.codebook.embed_avg', } lowercase__ = { 'encoder.model.0.conv.conv': 'encoder.layers.0.conv', 'encoder.model.1.block.1.conv.conv': 'encoder.layers.1.block.1.conv', 'encoder.model.1.block.3.conv.conv': 'encoder.layers.1.block.3.conv', 'encoder.model.1.shortcut.conv.conv': 'encoder.layers.1.shortcut.conv', 'encoder.model.3.conv.conv': 'encoder.layers.3.conv', 'encoder.model.4.block.1.conv.conv': 'encoder.layers.4.block.1.conv', 'encoder.model.4.block.3.conv.conv': 'encoder.layers.4.block.3.conv', 'encoder.model.4.shortcut.conv.conv': 'encoder.layers.4.shortcut.conv', 'encoder.model.6.conv.conv': 'encoder.layers.6.conv', 'encoder.model.7.block.1.conv.conv': 'encoder.layers.7.block.1.conv', 'encoder.model.7.block.3.conv.conv': 'encoder.layers.7.block.3.conv', 'encoder.model.7.shortcut.conv.conv': 'encoder.layers.7.shortcut.conv', 'encoder.model.9.conv.conv': 'encoder.layers.9.conv', 'encoder.model.10.block.1.conv.conv': 'encoder.layers.10.block.1.conv', 'encoder.model.10.block.3.conv.conv': 'encoder.layers.10.block.3.conv', 'encoder.model.10.shortcut.conv.conv': 'encoder.layers.10.shortcut.conv', 'encoder.model.12.conv.conv': 'encoder.layers.12.conv', 'encoder.model.13.lstm': 'encoder.layers.13.lstm', 'encoder.model.15.conv.conv': 'encoder.layers.15.conv', } lowercase__ = { 'encoder.model.0.conv.norm': 'encoder.layers.0.norm', 'encoder.model.1.block.1.conv.norm': 'encoder.layers.1.block.1.norm', 'encoder.model.1.block.3.conv.norm': 'encoder.layers.1.block.3.norm', 'encoder.model.1.shortcut.conv.norm': 'encoder.layers.1.shortcut.norm', 'encoder.model.3.conv.norm': 'encoder.layers.3.norm', 'encoder.model.4.block.1.conv.norm': 'encoder.layers.4.block.1.norm', 'encoder.model.4.block.3.conv.norm': 'encoder.layers.4.block.3.norm', 'encoder.model.4.shortcut.conv.norm': 'encoder.layers.4.shortcut.norm', 'encoder.model.6.conv.norm': 'encoder.layers.6.norm', 'encoder.model.7.block.1.conv.norm': 'encoder.layers.7.block.1.norm', 'encoder.model.7.block.3.conv.norm': 'encoder.layers.7.block.3.norm', 'encoder.model.7.shortcut.conv.norm': 'encoder.layers.7.shortcut.norm', 'encoder.model.9.conv.norm': 'encoder.layers.9.norm', 'encoder.model.10.block.1.conv.norm': 'encoder.layers.10.block.1.norm', 'encoder.model.10.block.3.conv.norm': 'encoder.layers.10.block.3.norm', 'encoder.model.10.shortcut.conv.norm': 'encoder.layers.10.shortcut.norm', 'encoder.model.12.conv.norm': 'encoder.layers.12.norm', 'encoder.model.15.conv.norm': 'encoder.layers.15.norm', } lowercase__ = { 'decoder.model.0.conv.conv': 'decoder.layers.0.conv', 'decoder.model.1.lstm': 'decoder.layers.1.lstm', 'decoder.model.3.convtr.convtr': 'decoder.layers.3.conv', 'decoder.model.4.block.1.conv.conv': 'decoder.layers.4.block.1.conv', 'decoder.model.4.block.3.conv.conv': 'decoder.layers.4.block.3.conv', 'decoder.model.4.shortcut.conv.conv': 'decoder.layers.4.shortcut.conv', 'decoder.model.6.convtr.convtr': 'decoder.layers.6.conv', 'decoder.model.7.block.1.conv.conv': 'decoder.layers.7.block.1.conv', 'decoder.model.7.block.3.conv.conv': 'decoder.layers.7.block.3.conv', 'decoder.model.7.shortcut.conv.conv': 'decoder.layers.7.shortcut.conv', 'decoder.model.9.convtr.convtr': 'decoder.layers.9.conv', 'decoder.model.10.block.1.conv.conv': 'decoder.layers.10.block.1.conv', 'decoder.model.10.block.3.conv.conv': 'decoder.layers.10.block.3.conv', 'decoder.model.10.shortcut.conv.conv': 'decoder.layers.10.shortcut.conv', 'decoder.model.12.convtr.convtr': 'decoder.layers.12.conv', 'decoder.model.13.block.1.conv.conv': 'decoder.layers.13.block.1.conv', 'decoder.model.13.block.3.conv.conv': 'decoder.layers.13.block.3.conv', 'decoder.model.13.shortcut.conv.conv': 'decoder.layers.13.shortcut.conv', 'decoder.model.15.conv.conv': 'decoder.layers.15.conv', } lowercase__ = { 'decoder.model.0.conv.norm': 'decoder.layers.0.norm', 'decoder.model.3.convtr.norm': 'decoder.layers.3.norm', 'decoder.model.4.block.1.conv.norm': 'decoder.layers.4.block.1.norm', 'decoder.model.4.block.3.conv.norm': 'decoder.layers.4.block.3.norm', 'decoder.model.4.shortcut.conv.norm': 'decoder.layers.4.shortcut.norm', 'decoder.model.6.convtr.norm': 'decoder.layers.6.norm', 'decoder.model.7.block.1.conv.norm': 'decoder.layers.7.block.1.norm', 'decoder.model.7.block.3.conv.norm': 'decoder.layers.7.block.3.norm', 'decoder.model.7.shortcut.conv.norm': 'decoder.layers.7.shortcut.norm', 'decoder.model.9.convtr.norm': 'decoder.layers.9.norm', 'decoder.model.10.block.1.conv.norm': 'decoder.layers.10.block.1.norm', 'decoder.model.10.block.3.conv.norm': 'decoder.layers.10.block.3.norm', 'decoder.model.10.shortcut.conv.norm': 'decoder.layers.10.shortcut.norm', 'decoder.model.12.convtr.norm': 'decoder.layers.12.norm', 'decoder.model.13.block.1.conv.norm': 'decoder.layers.13.block.1.norm', 'decoder.model.13.block.3.conv.norm': 'decoder.layers.13.block.3.norm', 'decoder.model.13.shortcut.conv.norm': 'decoder.layers.13.shortcut.norm', 'decoder.model.15.conv.norm': 'decoder.layers.15.norm', } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } lowercase__ = [] lowercase__ = [] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: for attribute in key.split('.' ): a__: str = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: str = value elif weight_type == "weight_g": a__: int = value elif weight_type == "weight_v": a__: Tuple = value elif weight_type == "bias": a__: Dict = value elif weight_type == "running_mean": a__: Any = value elif weight_type == "running_var": a__: Tuple = value elif weight_type == "num_batches_tracked": a__: List[str] = value elif weight_type == "weight_ih_l0": a__: List[Any] = value elif weight_type == "weight_hh_l0": a__: List[Any] = value elif weight_type == "bias_ih_l0": a__: List[Any] = value elif weight_type == "bias_hh_l0": a__: List[Any] = value elif weight_type == "weight_ih_l1": a__: int = value elif weight_type == "weight_hh_l1": a__: str = value elif weight_type == "bias_ih_l1": a__: Union[str, Any] = value elif weight_type == "bias_hh_l1": a__: Any = value else: a__: Union[str, Any] = value logger.info(F'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: for key in ignore_keys: if key.endswith('.*' ): if name.startswith(key[:-1] ): return True elif ".*." in key: a__ , a__: Optional[Any] = key.split('.*.' ) if prefix in name and suffix in name: return True elif key in name: return True return False def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[Any] = [] if model_name == "encodec_24khz" or "encodec_32khz": a__: Optional[int] = MAPPING_24K elif model_name == "encodec_48khz": a__: List[Any] = MAPPING_48K else: raise ValueError(F'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): logger.info(F'{name} was ignored' ) continue a__: int = False for key, mapped_key in MAPPING.items(): if "*" in key: a__ , a__: str = key.split('.*.' ) if prefix in name and suffix in name: a__: List[str] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('embed' ) and name.endswith('embed_avg' ): continue a__: List[str] = True if "*" in mapped_key: a__: List[str] = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: str = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: int = 'weight_g' elif "weight_v" in name: a__: Dict = 'weight_v' elif "weight_ih_l0" in name: a__: int = 'weight_ih_l0' elif "weight_hh_l0" in name: a__: Union[str, Any] = 'weight_hh_l0' elif "bias_ih_l0" in name: a__: Optional[Any] = 'bias_ih_l0' elif "bias_hh_l0" in name: a__: Optional[int] = 'bias_hh_l0' elif "weight_ih_l1" in name: a__: Dict = 'weight_ih_l1' elif "weight_hh_l1" in name: a__: Optional[Any] = 'weight_hh_l1' elif "bias_ih_l1" in name: a__: List[str] = 'bias_ih_l1' elif "bias_hh_l1" in name: a__: Optional[Any] = 'bias_hh_l1' elif "bias" in name: a__: List[str] = 'bias' elif "weight" in name: a__: Any = 'weight' elif "running_mean" in name: a__: Dict = 'running_mean' elif "running_var" in name: a__: Dict = 'running_var' elif "num_batches_tracked" in name: a__: Dict = 'num_batches_tracked' else: a__: List[str] = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->int: if config_path is not None: a__: Dict = EncodecConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: a__: Tuple = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": a__: Any = [8, 5, 4, 4] a__: List[str] = [2.2] a__: List[Any] = 64 a__: Dict = 32000 a__: Union[str, Any] = 2048 a__: Union[str, Any] = False a__: Any = False a__: Optional[Any] = False elif model_name == "encodec_48khz": a__: Optional[int] = [8, 5, 4, 2] a__: Union[str, Any] = [3.0, 6.0, 12.0, 24.0] a__: List[str] = 48000 a__: Tuple = 2 a__: Optional[Any] = False a__: Optional[int] = 'time_group_norm' a__: Union[str, Any] = True a__: Dict = 1.0 a__: str = 0.01 else: raise ValueError(F'Unknown model name: {model_name}' ) a__: Optional[int] = EncodecModel(_SCREAMING_SNAKE_CASE ) a__: List[str] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: int = torch.load(_SCREAMING_SNAKE_CASE ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights a__: str = original_checkpoint['best_state'] recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if repo_id: print('Pushing to the hub...' ) feature_extractor.push_to_hub(_SCREAMING_SNAKE_CASE ) model.push_to_hub(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
290
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'alibaba-damo/mgp-str-base': 'https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json', } class __snake_case ( __lowerCAmelCase ): a__ = """mgp-str""" def __init__( self , lowercase=[32, 1_28] , lowercase=4 , lowercase=3 , lowercase=27 , lowercase=38 , lowercase=5_02_57 , lowercase=3_05_22 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=4.0 , lowercase=True , lowercase=False , lowercase=1e-5 , lowercase=0.0 , lowercase=0.0 , lowercase=0.0 , lowercase=False , lowercase=0.02 , **lowercase , ) -> Optional[int]: '''simple docstring''' super().__init__(**lowercase) a__: Optional[int] = image_size a__: Dict = patch_size a__: Union[str, Any] = num_channels a__: Tuple = max_token_length a__: Optional[Any] = num_character_labels a__: List[str] = num_bpe_labels a__: List[Any] = num_wordpiece_labels a__: Tuple = hidden_size a__: str = num_hidden_layers a__: str = num_attention_heads a__: int = mlp_ratio a__: Tuple = distilled a__: List[Any] = layer_norm_eps a__: Dict = drop_rate a__: List[str] = qkv_bias a__: str = attn_drop_rate a__: Any = drop_path_rate a__: Union[str, Any] = output_aa_attentions a__: Dict = initializer_range
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: print('moving disk from' , _SCREAMING_SNAKE_CASE , 'to' , _SCREAMING_SNAKE_CASE ) def __a ( ) ->List[str]: a__: Dict = int(input('Height of hanoi: ' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
290
1
"""simple docstring""" from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class __snake_case ( nn.Module ): def __init__( self , lowercase = 16 , lowercase = 88 , lowercase = None , lowercase = 1 , lowercase = 0.0 , lowercase = 32 , lowercase = None , lowercase = False , lowercase = None , lowercase = None , lowercase = "geglu" , lowercase = None , ) -> str: '''simple docstring''' super().__init__() a__: str = nn.ModuleList( [ TransformeraDModel( num_attention_heads=lowercase , attention_head_dim=lowercase , in_channels=lowercase , num_layers=lowercase , dropout=lowercase , norm_num_groups=lowercase , cross_attention_dim=lowercase , attention_bias=lowercase , sample_size=lowercase , num_vector_embeds=lowercase , activation_fn=lowercase , num_embeds_ada_norm=lowercase , ) for _ in range(2) ]) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference a__: List[str] = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` a__: Union[str, Any] = [77, 2_57] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` a__: Any = [1, 0] def lowerCamelCase_ ( self , lowercase , lowercase , lowercase=None , lowercase=None , lowercase=None , lowercase = True , ) -> List[Any]: '''simple docstring''' a__: int = hidden_states a__: List[str] = [] a__: Optional[Any] = 0 # attention_mask is not used yet for i in range(2): # for each of the two transformers, pass the corresponding condition tokens a__: Tuple = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] a__: Optional[int] = self.transformer_index_for_condition[i] a__: List[str] = self.transformers[transformer_index]( lowercase , encoder_hidden_states=lowercase , timestep=lowercase , cross_attention_kwargs=lowercase , return_dict=lowercase , )[0] encoded_states.append(encoded_state - input_states) tokens_start += self.condition_lengths[i] a__: Union[str, Any] = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) a__: Any = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=lowercase)
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = False ) ->str: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = F'Expected string as input, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[str] = F'Expected boolean as use_pascal parameter, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) a__: int = input_str.split('_' ) a__: List[str] = 0 if use_pascal else 1 a__: List[str] = words[start_index:] a__: List[str] = [word[0].upper() + word[1:] for word in words_to_capitalize] a__: List[str] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = { 'configuration_bigbird_pegasus': [ 'BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BigBirdPegasusConfig', 'BigBirdPegasusOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST', 'BigBirdPegasusForCausalLM', 'BigBirdPegasusForConditionalGeneration', 'BigBirdPegasusForQuestionAnswering', 'BigBirdPegasusForSequenceClassification', 'BigBirdPegasusModel', 'BigBirdPegasusPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" class __snake_case : def __init__( self , lowercase , lowercase=None , lowercase=None) -> List[str]: '''simple docstring''' a__: Dict = data a__: List[Any] = previous a__: Any = next_node def __str__( self) -> str: '''simple docstring''' return f'{self.data}' def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.data def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.next def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return self.previous class __snake_case : def __init__( self , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = head def __iter__( self) -> List[Any]: '''simple docstring''' return self def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if not self.current: raise StopIteration else: a__: Dict = self.current.get_data() a__: Optional[Any] = self.current.get_next() return value class __snake_case : def __init__( self) -> Dict: '''simple docstring''' a__: List[Any] = None # First node in list a__: Optional[int] = None # Last node in list def __str__( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.head a__: Optional[Any] = [] while current is not None: nodes.append(current.get_data()) a__: str = current.get_next() return " ".join(str(lowercase) for node in nodes) def __contains__( self , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = self.head while current: if current.get_data() == value: return True a__: Dict = current.get_next() return False def __iter__( self) -> int: '''simple docstring''' return LinkedListIterator(self.head) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: a__: Optional[Any] = node a__: Optional[Any] = node else: self.insert_before_node(self.head , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: self.set_head(lowercase) else: self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = Node(lowercase) if self.head is None: self.set_head(lowercase) else: self.set_tail(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Union[str, Any] = node a__: Optional[Any] = node.previous if node.get_previous() is None: a__: Tuple = node_to_insert else: a__: int = node_to_insert a__: Optional[int] = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Optional[int] = node a__: Tuple = node.next if node.get_next() is None: a__: Optional[int] = node_to_insert else: a__: Any = node_to_insert a__: str = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Any = 1 a__: Tuple = Node(lowercase) a__: Tuple = self.head while node: if current_position == position: self.insert_before_node(lowercase , lowercase) return current_position += 1 a__: List[Any] = node.next self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> Node: '''simple docstring''' a__: Tuple = self.head while node: if node.get_data() == item: return node a__: List[str] = node.get_next() raise Exception('Node not found') def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' if (node := self.get_node(lowercase)) is not None: if node == self.head: a__: Any = self.head.get_next() if node == self.tail: a__: List[Any] = self.tail.get_previous() self.remove_node_pointers(lowercase) @staticmethod def lowerCamelCase_ ( lowercase) -> None: '''simple docstring''' if node.get_next(): a__: Any = node.previous if node.get_previous(): a__: List[str] = node.next a__: int = None a__: Union[str, Any] = None def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self.head is None def __a ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class __snake_case ( __lowerCAmelCase ): a__ = """yolos""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=[5_12, 8_64] , lowercase=16 , lowercase=3 , lowercase=True , lowercase=1_00 , lowercase=True , lowercase=False , lowercase=1 , lowercase=5 , lowercase=2 , lowercase=5 , lowercase=2 , lowercase=0.1 , **lowercase , ) -> List[Any]: '''simple docstring''' super().__init__(**lowercase) a__: Optional[int] = hidden_size a__: List[str] = num_hidden_layers a__: Optional[Any] = num_attention_heads a__: List[str] = intermediate_size a__: Tuple = hidden_act a__: Any = hidden_dropout_prob a__: Dict = attention_probs_dropout_prob a__: List[Any] = initializer_range a__: int = layer_norm_eps a__: str = image_size a__: Optional[Any] = patch_size a__: Dict = num_channels a__: Optional[Any] = qkv_bias a__: List[str] = num_detection_tokens a__: Tuple = use_mid_position_embeddings a__: Tuple = auxiliary_loss # Hungarian matcher a__: Optional[int] = class_cost a__: Tuple = bbox_cost a__: Optional[int] = giou_cost # Loss coefficients a__: Union[str, Any] = bbox_loss_coefficient a__: List[str] = giou_loss_coefficient a__: Optional[int] = eos_coefficient class __snake_case ( __lowerCAmelCase ): a__ = version.parse("""1.11""" ) @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ]) @property def lowerCamelCase_ ( self) -> float: '''simple docstring''' return 1e-4 @property def lowerCamelCase_ ( self) -> int: '''simple docstring''' return 12
290
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
1
"""simple docstring""" lowercase__ = { 'meter': 'm', 'kilometer': 'km', 'megametre': 'Mm', 'gigametre': 'Gm', 'terametre': 'Tm', 'petametre': 'Pm', 'exametre': 'Em', 'zettametre': 'Zm', 'yottametre': 'Ym', } # Exponent of the factor(meter) lowercase__ = { 'm': 0, 'km': 3, 'Mm': 6, 'Gm': 9, 'Tm': 12, 'Pm': 15, 'Em': 18, 'Zm': 21, 'Ym': 24, } def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: a__: List[Any] = from_type.lower().strip('s' ) a__: Tuple = to_type.lower().strip('s' ) a__: Tuple = UNIT_SYMBOL.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Dict = UNIT_SYMBOL.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if from_sanitized not in METRIC_CONVERSION: a__: Any = ( F'Invalid \'from_type\' value: {from_type!r}.\n' F'Conversion abbreviations are: {", ".join(_SCREAMING_SNAKE_CASE )}' ) raise ValueError(_SCREAMING_SNAKE_CASE ) if to_sanitized not in METRIC_CONVERSION: a__: Dict = ( F'Invalid \'to_type\' value: {to_type!r}.\n' F'Conversion abbreviations are: {", ".join(_SCREAMING_SNAKE_CASE )}' ) raise ValueError(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = METRIC_CONVERSION[from_sanitized] a__: Union[str, Any] = METRIC_CONVERSION[to_sanitized] a__: List[Any] = 1 if from_exponent > to_exponent: a__: Tuple = from_exponent - to_exponent else: a__: List[str] = -(to_exponent - from_exponent) return value * pow(10 , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": from doctest import testmod testmod()
290
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowercase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Any: if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): a__: Optional[int] = [image] a__: str = [trans(img.convert('RGB' ) ) for img in image] a__: Any = torch.stack(_SCREAMING_SNAKE_CASE ) return image class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM a__: Dict = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowercase , scheduler=lowercase) def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f'The value of strength should in [0.0, 1.0] but is {strength}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: int = min(int(num_inference_steps * strength) , lowercase) a__: Any = max(num_inference_steps - init_timestep , 0) a__: Union[str, Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> List[Any]: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( f'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase)}') a__: Tuple = image.to(device=lowercase , dtype=lowercase) if isinstance(lowercase , lowercase) and len(lowercase) != batch_size: raise ValueError( f'You have passed a list of generators of length {len(lowercase)}, but requested an effective batch' f' size of {batch_size}. Make sure the batch size matches the length of the generators.') a__: List[str] = init_latents.shape a__: List[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) # get latents print('add noise to latents at timestep' , lowercase) a__: int = self.scheduler.add_noise(lowercase , lowercase , lowercase) a__: Dict = init_latents return latents @torch.no_grad() def __call__( self , lowercase = None , lowercase = 0.8 , lowercase = 1 , lowercase = None , lowercase = 0.0 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase) # 2. Preprocess image a__: Tuple = preprocess(lowercase) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device) a__ , a__: Union[str, Any] = self.get_timesteps(lowercase , lowercase , self.device) a__: Optional[int] = timesteps[:1].repeat(lowercase) # 4. Prepare latent variables a__: Union[str, Any] = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase) a__: Optional[Any] = latents # 5. Denoising loop for t in self.progress_bar(lowercase): # 1. predict noise model_output a__: Dict = self.unet(lowercase , lowercase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample a__: Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1) a__: Optional[int] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": a__: Dict = self.numpy_to_pil(lowercase) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase)
290
1
"""simple docstring""" lowercase__ = {str(digit): digit**5 for digit in range(10)} def __a ( _SCREAMING_SNAKE_CASE ) ->int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(_SCREAMING_SNAKE_CASE ) ) def __a ( ) ->int: return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": print(solution())
290
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: Optional[int] = SamImageProcessor() a__: Tuple = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> List[Any]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[Any] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Optional[int] = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: List[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = self.get_image_processor() a__: List[Any] = SamProcessor(image_processor=lowercase) a__: Optional[int] = self.prepare_image_inputs() a__: Optional[Any] = image_processor(lowercase , return_tensors='np') a__: Tuple = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_torch def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.get_image_processor() a__: List[str] = SamProcessor(image_processor=lowercase) a__: Optional[Any] = [torch.ones((1, 3, 5, 5))] a__: Union[str, Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: int = processor.post_process_masks(lowercase , lowercase , lowercase) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Optional[int] = processor.post_process_masks( lowercase , torch.tensor(lowercase) , torch.tensor(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Dict = [np.ones((1, 3, 5, 5))] a__: Tuple = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = [[1, 0], [0, 1]] with self.assertRaises(lowercase): a__: List[Any] = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) @require_vision @require_tf class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: List[Any] = SamImageProcessor() a__: Optional[int] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> int: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[int] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Dict = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = self.get_image_processor() a__: str = SamProcessor(image_processor=lowercase) a__: int = self.prepare_image_inputs() a__: int = image_processor(lowercase , return_tensors='np') a__: Dict = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_tf def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Any = SamProcessor(image_processor=lowercase) a__: str = [tf.ones((1, 3, 5, 5))] a__: List[Any] = [[17_64, 26_46]] a__: List[Any] = [[6_83, 10_24]] a__: List[Any] = processor.post_process_masks(lowercase , lowercase , lowercase , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = processor.post_process_masks( lowercase , tf.convert_to_tensor(lowercase) , tf.convert_to_tensor(lowercase) , return_tensors='tf' , ) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Optional[Any] = [np.ones((1, 3, 5, 5))] a__: int = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: List[str] = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError): a__: Any = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = tempfile.mkdtemp() a__: int = SamImageProcessor() a__: Union[str, Any] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Any = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.get_image_processor() a__: int = SamProcessor(image_processor=lowercase) a__: int = np.random.randint(0 , 2 , size=(1, 3, 5, 5)).astype(np.floataa) a__: Dict = [tf.convert_to_tensor(lowercase)] a__: Union[str, Any] = [torch.tensor(lowercase)] a__: List[Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: Tuple = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='tf') a__: str = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='pt') self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy())) @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Dict = SamProcessor(image_processor=lowercase) a__: Any = self.prepare_image_inputs() a__: List[Any] = image_processor(lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Tuple = processor(images=lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Any = image_processor(lowercase , return_tensors='tf')['pixel_values'].numpy() a__: Any = processor(images=lowercase , return_tensors='tf')['pixel_values'].numpy() self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase))
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: print('moving disk from' , _SCREAMING_SNAKE_CASE , 'to' , _SCREAMING_SNAKE_CASE ) def __a ( ) ->List[str]: a__: Dict = int(input('Height of hanoi: ' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
290
"""simple docstring""" from math import pow, sqrt def __a ( *_SCREAMING_SNAKE_CASE ) ->bool: a__: Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) > 0 and all(value > 0.0 for value in values ) return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
1
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase=7_68) -> str: '''simple docstring''' super().__init__(lowercase) a__: Dict = proj_size a__: List[str] = CLIPVisionModel(lowercase) a__: Tuple = PaintByExampleMapper(lowercase) a__: List[str] = nn.LayerNorm(config.hidden_size) a__: Optional[Any] = nn.Linear(config.hidden_size , self.proj_size) # uncondition for scaling a__: List[Any] = nn.Parameter(torch.randn((1, 1, self.proj_size))) def lowerCamelCase_ ( self , lowercase , lowercase=False) -> Tuple: '''simple docstring''' a__: str = self.model(pixel_values=lowercase) a__: Union[str, Any] = clip_output.pooler_output a__: int = self.mapper(latent_states[:, None]) a__: int = self.final_layer_norm(lowercase) a__: Optional[Any] = self.proj_out(lowercase) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class __snake_case ( nn.Module ): def __init__( self , lowercase) -> Dict: '''simple docstring''' super().__init__() a__: Optional[Any] = (config.num_hidden_layers + 1) // 5 a__: Optional[Any] = config.hidden_size a__: int = 1 a__: Optional[Any] = nn.ModuleList( [ BasicTransformerBlock(lowercase , lowercase , lowercase , activation_fn='gelu' , attention_bias=lowercase) for _ in range(lowercase) ]) def lowerCamelCase_ ( self , lowercase) -> Tuple: '''simple docstring''' for block in self.blocks: a__: str = block(lowercase) return hidden_states
290
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """roberta-prelayernorm""" def __init__( self , lowercase=5_02_65 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=True , lowercase=None , **lowercase , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase) a__: Union[str, Any] = vocab_size a__: str = hidden_size a__: Tuple = num_hidden_layers a__: List[str] = num_attention_heads a__: Dict = hidden_act a__: int = intermediate_size a__: Tuple = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: Tuple = max_position_embeddings a__: Tuple = type_vocab_size a__: Optional[Any] = initializer_range a__: Tuple = layer_norm_eps a__: Optional[int] = position_embedding_type a__: Any = use_cache a__: Dict = classifier_dropout class __snake_case ( __lowerCAmelCase ): @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": a__: str = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a__: Union[str, Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
290
1
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """audio-spectrogram-transformer""" def __init__( self , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=1e-12 , lowercase=16 , lowercase=True , lowercase=10 , lowercase=10 , lowercase=10_24 , lowercase=1_28 , **lowercase , ) -> str: '''simple docstring''' super().__init__(**lowercase) a__: Any = hidden_size a__: int = num_hidden_layers a__: Union[str, Any] = num_attention_heads a__: Any = intermediate_size a__: Union[str, Any] = hidden_act a__: int = hidden_dropout_prob a__: str = attention_probs_dropout_prob a__: str = initializer_range a__: Tuple = layer_norm_eps a__: Any = patch_size a__: int = qkv_bias a__: Optional[Any] = frequency_stride a__: int = time_stride a__: List[str] = max_length a__: Tuple = num_mel_bins
290
1
"""simple docstring""" from math import pi, sqrt, tan def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) a__: List[Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(_SCREAMING_SNAKE_CASE , 2 ) * torus_radius * tube_radius def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) a__: int = (sidea + sidea + sidea) / 2 a__: Tuple = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(f"Rectangle: {area_rectangle(10, 20) = }") print(f"Square: {area_square(10) = }") print(f"Triangle: {area_triangle(10, 10) = }") print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(f"Parallelogram: {area_parallelogram(10, 20) = }") print(f"Rhombus: {area_rhombus(10, 20) = }") print(f"Trapezium: {area_trapezium(10, 20, 30) = }") print(f"Circle: {area_circle(20) = }") print(f"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(f"Cube: {surface_area_cube(20) = }") print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(f"Sphere: {surface_area_sphere(20) = }") print(f"Hemisphere: {surface_area_hemisphere(20) = }") print(f"Cone: {surface_area_cone(10, 20) = }") print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(f"Cylinder: {surface_area_cylinder(10, 20) = }") print(f"Torus: {surface_area_torus(20, 10) = }") print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(f"Square: {area_reg_polygon(4, 10) = }") print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
290
"""simple docstring""" import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowercase__ = get_tests_dir('fixtures/test_sentencepiece.model') lowercase__ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') lowercase__ = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = CamembertTokenizer a__ = CamembertTokenizerFast a__ = True a__ = True def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing a__: Tuple = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Optional[Any] = '<pad>' a__: List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase) , lowercase) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase) , lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: str = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>NOTUSED') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(lowercase) , 10_04) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 10_05) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Optional[Any] = CamembertTokenizer(lowercase) tokenizer.save_pretrained(self.tmpdirname) a__: List[Any] = CamembertTokenizerFast.from_pretrained(self.tmpdirname) a__: Dict = 'I was born in 92000, and this is falsé.' a__: Optional[int] = tokenizer.encode(lowercase) a__: Any = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) a__: Optional[Any] = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) a__: Tuple = tokenizer.convert_ids_to_tokens(lowercase) a__: Tuple = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' if not self.test_rust_tokenizer: return a__: Dict = self.get_tokenizer() a__: str = self.get_rust_tokenizer() a__: int = 'I was born in 92000, and this is falsé.' a__: Optional[Any] = tokenizer.tokenize(lowercase) a__: List[Any] = rust_tokenizer.tokenize(lowercase) self.assertListEqual(lowercase , lowercase) a__: str = tokenizer.encode(lowercase , add_special_tokens=lowercase) a__: str = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase) self.assertListEqual(lowercase , lowercase) a__: Tuple = self.get_rust_tokenizer() a__: Union[str, Any] = tokenizer.encode(lowercase) a__: List[Any] = rust_tokenizer.encode(lowercase) self.assertListEqual(lowercase , lowercase) @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Union[str, Any] = {'input_ids': [[5, 54, 71_96, 2_97, 30, 23, 7_76, 18, 11, 32_15, 37_05, 82_52, 22, 31_64, 11_81, 21_16, 29, 16, 8_13, 25, 7_91, 33_14, 20, 34_46, 38, 2_75_75, 1_20, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 4_68, 17, 11, 90_88, 20, 15_17, 8, 2_28_04, 1_88_18, 10, 38, 6_29, 6_07, 6_07, 1_42, 19, 71_96, 8_67, 56, 1_03_26, 24, 22_67, 20, 4_16, 50_72, 1_56_12, 2_33, 7_34, 7, 23_99, 27, 16, 30_15, 16_49, 7, 24, 20, 43_38, 23_99, 27, 13, 34_00, 14, 13, 61_89, 8, 9_30, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. a__: int = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=lowercase , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=lowercase , )
290
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE = 1000000 ) ->int: a__: int = limit + 1 a__: Optional[int] = [0] * limit for first_term in range(1 , _SCREAMING_SNAKE_CASE ): for n in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a a__: Any = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f"{solution() = }")
290
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = ShapEImgaImgPipeline a__ = ["""image"""] a__ = ["""image"""] a__ = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] a__ = False @property def lowerCamelCase_ ( self) -> str: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Any: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 8 @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' torch.manual_seed(0) a__: str = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=64 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) a__: List[str] = CLIPVisionModel(lowercase) return model @property def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Optional[Any] = CLIPImageProcessor( crop_size=2_24 , do_center_crop=lowercase , do_normalize=lowercase , do_resize=lowercase , image_mean=[0.48145466, 0.4578275, 0.40821073] , image_std=[0.26862954, 0.26130258, 0.27577711] , resample=3 , size=2_24 , ) return image_processor @property def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0) a__: Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 16, 'embedding_dim': self.time_input_dim, 'num_embeddings': 32, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'embedding_proj_norm_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } a__: Optional[Any] = PriorTransformer(**lowercase) return model @property def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0) a__: List[Any] = { 'param_shapes': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 12, 'background': ( 0.1, 0.1, 0.1, ), } a__: Optional[Any] = ShapERenderer(**lowercase) return model def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = self.dummy_prior a__: Optional[Any] = self.dummy_image_encoder a__: str = self.dummy_image_processor a__: Tuple = self.dummy_renderer a__: int = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=10_24 , prediction_type='sample' , use_karras_sigmas=lowercase , clip_sample=lowercase , clip_sample_range=1.0 , ) a__: str = { 'prior': prior, 'image_encoder': image_encoder, 'image_processor': image_processor, 'renderer': renderer, 'scheduler': scheduler, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> int: '''simple docstring''' a__: Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) if str(lowercase).startswith('mps'): a__: Any = torch.manual_seed(lowercase) else: a__: Optional[Any] = torch.Generator(device=lowercase).manual_seed(lowercase) a__: Dict = { 'image': input_image, 'generator': generator, 'num_inference_steps': 1, 'frame_size': 32, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Dict = 'cpu' a__: Dict = self.get_dummy_components() a__: str = self.pipeline_class(**lowercase) a__: Tuple = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Any = pipe(**self.get_dummy_inputs(lowercase)) a__: List[str] = output.images[0] a__: Tuple = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) a__: int = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: str = torch_device == 'cpu' a__: Dict = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=lowercase , relax_max_difference=lowercase , ) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Tuple = self.get_dummy_components() a__: List[Any] = self.pipeline_class(**lowercase) a__: List[str] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Dict = 1 a__: Tuple = 2 a__: Tuple = self.get_dummy_inputs(lowercase) for key in inputs.keys(): if key in self.batch_params: a__: str = batch_size * [inputs[key]] a__: List[str] = pipe(**lowercase , num_images_per_prompt=lowercase)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/corgi.png') a__: Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_img2img_out.npy') a__: List[str] = ShapEImgaImgPipeline.from_pretrained('openai/shap-e-img2img') a__: List[str] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Tuple = torch.Generator(device=lowercase).manual_seed(0) a__: Optional[Any] = pipe( lowercase , generator=lowercase , guidance_scale=3.0 , num_inference_steps=64 , frame_size=64 , output_type='np' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(lowercase , lowercase)
290
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union lowercase__ = TypeVar('T') lowercase__ = Union[List[T], Tuple[T, ...]] lowercase__ = Union[T, List[T], Dict[str, T]] lowercase__ = Union[str, bytes, os.PathLike]
290
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'kssteven/ibert-roberta-base': 'https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json', 'kssteven/ibert-roberta-large': 'https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json', 'kssteven/ibert-roberta-large-mnli': ( 'https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """ibert""" def __init__( self , lowercase=3_05_22 , lowercase=7_68 , lowercase=12 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=1 , lowercase=0 , lowercase=2 , lowercase="absolute" , lowercase=False , lowercase="none" , **lowercase , ) -> Tuple: '''simple docstring''' super().__init__(pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , **lowercase) a__: Union[str, Any] = vocab_size a__: Dict = hidden_size a__: Optional[int] = num_hidden_layers a__: str = num_attention_heads a__: List[str] = hidden_act a__: int = intermediate_size a__: str = hidden_dropout_prob a__: Optional[Any] = attention_probs_dropout_prob a__: Union[str, Any] = max_position_embeddings a__: List[Any] = type_vocab_size a__: List[str] = initializer_range a__: Any = layer_norm_eps a__: Tuple = position_embedding_type a__: Union[str, Any] = quant_mode a__: List[str] = force_dequant class __snake_case ( __lowerCAmelCase ): @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": a__: List[str] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: a__: List[str] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ])
290
"""simple docstring""" from math import pi, sqrt, tan def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) a__: List[Any] = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(_SCREAMING_SNAKE_CASE , 2 ) * torus_radius * tube_radius def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def __a ( _SCREAMING_SNAKE_CASE ) ->float: if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) a__: int = (sidea + sidea + sidea) / 2 a__: Tuple = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def __a ( _SCREAMING_SNAKE_CASE ) ->float: if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('[DEMO] Areas of various geometric shapes: \n') print(f"Rectangle: {area_rectangle(10, 20) = }") print(f"Square: {area_square(10) = }") print(f"Triangle: {area_triangle(10, 10) = }") print(f"Triangle: {area_triangle_three_sides(5, 12, 13) = }") print(f"Parallelogram: {area_parallelogram(10, 20) = }") print(f"Rhombus: {area_rhombus(10, 20) = }") print(f"Trapezium: {area_trapezium(10, 20, 30) = }") print(f"Circle: {area_circle(20) = }") print(f"Ellipse: {area_ellipse(10, 20) = }") print('\nSurface Areas of various geometric shapes: \n') print(f"Cube: {surface_area_cube(20) = }") print(f"Cuboid: {surface_area_cuboid(10, 20, 30) = }") print(f"Sphere: {surface_area_sphere(20) = }") print(f"Hemisphere: {surface_area_hemisphere(20) = }") print(f"Cone: {surface_area_cone(10, 20) = }") print(f"Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }") print(f"Cylinder: {surface_area_cylinder(10, 20) = }") print(f"Torus: {surface_area_torus(20, 10) = }") print(f"Equilateral Triangle: {area_reg_polygon(3, 10) = }") print(f"Square: {area_reg_polygon(4, 10) = }") print(f"Reqular Pentagon: {area_reg_polygon(5, 10) = }")
290
1
"""simple docstring""" import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: List[str] = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: a__: Union[str, Any] = 128 elif "12-12" in model_name: a__: List[Any] = 12 a__: str = 12 elif "14-14" in model_name: a__: Dict = 14 a__: List[Any] = 14 elif "16-16" in model_name: a__: Any = 16 a__: List[str] = 16 else: raise ValueError('Model not supported' ) a__: int = 'huggingface/label-files' if "speech-commands" in model_name: a__: Tuple = 35 a__: str = 'speech-commands-v2-id2label.json' else: a__: Union[str, Any] = 527 a__: int = 'audioset-id2label.json' a__: Optional[Any] = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) a__: int = {int(_SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} a__: List[str] = idalabel a__: List[Any] = {v: k for k, v in idalabel.items()} return config def __a ( _SCREAMING_SNAKE_CASE ) ->Dict: if "module.v" in name: a__: Tuple = name.replace('module.v' , 'audio_spectrogram_transformer' ) if "cls_token" in name: a__: List[Any] = name.replace('cls_token' , 'embeddings.cls_token' ) if "dist_token" in name: a__: Optional[int] = name.replace('dist_token' , 'embeddings.distillation_token' ) if "pos_embed" in name: a__: List[Any] = name.replace('pos_embed' , 'embeddings.position_embeddings' ) if "patch_embed.proj" in name: a__: str = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) # transformer blocks if "blocks" in name: a__: Dict = name.replace('blocks' , 'encoder.layer' ) if "attn.proj" in name: a__: List[str] = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: a__: int = name.replace('attn' , 'attention.self' ) if "norm1" in name: a__: int = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: a__: Optional[Any] = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: a__: Tuple = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: a__: Tuple = name.replace('mlp.fc2' , 'output.dense' ) # final layernorm if "audio_spectrogram_transformer.norm" in name: a__: Optional[Any] = name.replace('audio_spectrogram_transformer.norm' , 'audio_spectrogram_transformer.layernorm' ) # classifier head if "module.mlp_head.0" in name: a__: Optional[Any] = name.replace('module.mlp_head.0' , 'classifier.layernorm' ) if "module.mlp_head.1" in name: a__: Tuple = name.replace('module.mlp_head.1' , 'classifier.dense' ) return name def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Any: for key in orig_state_dict.copy().keys(): a__: Optional[Any] = orig_state_dict.pop(_SCREAMING_SNAKE_CASE ) if "qkv" in key: a__: Optional[int] = key.split('.' ) a__: Dict = int(key_split[3] ) a__: Any = config.hidden_size if "weight" in key: a__: Any = val[:dim, :] a__: Any = val[dim : dim * 2, :] a__: str = val[-dim:, :] else: a__: Optional[Any] = val[:dim] a__: Dict = val[dim : dim * 2] a__: Optional[int] = val[-dim:] else: a__: Union[str, Any] = val return orig_state_dict def __a ( _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: Optional[Any] = [ 'module.v.head.weight', 'module.v.head.bias', 'module.v.head_dist.weight', 'module.v.head_dist.bias', ] for k in ignore_keys: state_dict.pop(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) ->int: a__: Tuple = get_audio_spectrogram_transformer_config(_SCREAMING_SNAKE_CASE ) a__: List[Any] = { 'ast-finetuned-audioset-10-10-0.4593': ( 'https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1' ), 'ast-finetuned-audioset-10-10-0.450': ( 'https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1' ), 'ast-finetuned-audioset-10-10-0.448': ( 'https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1' ), 'ast-finetuned-audioset-10-10-0.448-v2': ( 'https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1' ), 'ast-finetuned-audioset-12-12-0.447': ( 'https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1' ), 'ast-finetuned-audioset-14-14-0.443': ( 'https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1' ), 'ast-finetuned-audioset-16-16-0.442': ( 'https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1' ), 'ast-finetuned-speech-commands-v2': ( 'https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1' ), } # load original state_dict a__: int = model_name_to_url[model_name] a__: str = torch.hub.load_state_dict_from_url(_SCREAMING_SNAKE_CASE , map_location='cpu' ) # remove some keys remove_keys(_SCREAMING_SNAKE_CASE ) # rename some keys a__: Optional[int] = convert_state_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # load 🤗 model a__: Optional[int] = ASTForAudioClassification(_SCREAMING_SNAKE_CASE ) model.eval() model.load_state_dict(_SCREAMING_SNAKE_CASE ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 a__: Any = -4.2_677_393 if 'speech-commands' not in model_name else -6.845_978 a__: Dict = 4.5_689_974 if 'speech-commands' not in model_name else 5.5_654_526 a__: List[Any] = 1024 if 'speech-commands' not in model_name else 128 a__: str = ASTFeatureExtractor(mean=_SCREAMING_SNAKE_CASE , std=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE ) if "speech-commands" in model_name: a__: Tuple = load_dataset('speech_commands' , 'v0.02' , split='validation' ) a__: Optional[int] = dataset[0]['audio']['array'] else: a__: Optional[Any] = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' , ) a__ , a__: List[str] = torchaudio.load(_SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = waveform.squeeze().numpy() a__: Tuple = feature_extractor(_SCREAMING_SNAKE_CASE , sampling_rate=16000 , return_tensors='pt' ) # forward pass a__: Optional[int] = model(**_SCREAMING_SNAKE_CASE ) a__: Tuple = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": a__: str = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": a__: str = torch.tensor([-1.1_986, -7.0_903, -8.2_718] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": a__: Tuple = torch.tensor([-2.6_128, -8.0_080, -9.4_344] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": a__: Tuple = torch.tensor([-1.5_080, -7.4_534, -8.8_917] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": a__: Dict = torch.tensor([-0.5_050, -6.5_833, -8.0_843] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": a__: Optional[Any] = torch.tensor([-0.3_826, -7.0_336, -8.2_413] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": a__: Optional[Any] = torch.tensor([-1.2_113, -6.9_101, -8.3_470] ) elif model_name == "ast-finetuned-speech-commands-v2": a__: Tuple = torch.tensor([6.1_589, -8.0_566, -8.7_984] ) else: raise ValueError('Unknown model name' ) if not torch.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ): raise ValueError('Logits don\'t match' ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) print(F'Saving feature extractor to {pytorch_dump_folder_path}' ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) if push_to_hub: print('Pushing model and feature extractor to the hub...' ) model.push_to_hub(F'MIT/{model_name}' ) feature_extractor.push_to_hub(F'MIT/{model_name}' ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='ast-finetuned-audioset-10-10-0.4593', type=str, help='Name of the Audio Spectrogram Transformer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
290
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin lowercase__ = random.Random() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: if rng is None: a__: Any = global_rng a__: int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=4_00 , lowercase=20_00 , lowercase=1 , lowercase=0.0 , lowercase=1_60_00 , lowercase=True , lowercase=True , ) -> Union[str, Any]: '''simple docstring''' a__: Tuple = parent a__: Optional[int] = batch_size a__: Optional[Any] = min_seq_length a__: Optional[int] = max_seq_length a__: Tuple = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a__: Dict = feature_size a__: Any = padding_value a__: Optional[Any] = sampling_rate a__: Optional[Any] = return_attention_mask a__: str = do_normalize def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def lowerCamelCase_ ( self , lowercase=False , lowercase=False) -> Tuple: '''simple docstring''' def _flatten(lowercase): return list(itertools.chain(*lowercase)) if equal_length: a__: Dict = floats_list((self.batch_size, self.max_seq_length)) else: # make sure that inputs increase in size a__: List[Any] = [ _flatten(floats_list((x, self.feature_size))) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: a__: str = [np.asarray(lowercase) for x in speech_inputs] return speech_inputs class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = WavaVecaFeatureExtractor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = WavaVecaFeatureExtractionTester(self) def lowerCamelCase_ ( self , lowercase) -> List[Any]: '''simple docstring''' self.assertTrue(np.all(np.mean(lowercase , axis=0) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(lowercase , axis=0) - 1) < 1e-3)) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 a__: Optional[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: List[str] = [np.asarray(lowercase) for speech_input in speech_inputs] # Test not batched input a__: Optional[Any] = feat_extract(speech_inputs[0] , return_tensors='np').input_values a__: Dict = feat_extract(np_speech_inputs[0] , return_tensors='np').input_values self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test batched a__: Dict = feat_extract(lowercase , return_tensors='np').input_values a__: int = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) # Test 2-D numpy arrays are batched. a__: int = [floats_list((1, x))[0] for x in (8_00, 8_00, 8_00)] a__: Union[str, Any] = np.asarray(lowercase) a__: int = feat_extract(lowercase , return_tensors='np').input_values a__: Any = feat_extract(lowercase , return_tensors='np').input_values for enc_seq_a, enc_seq_a in zip(lowercase , lowercase): self.assertTrue(np.allclose(lowercase , lowercase , atol=1e-3)) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: List[Any] = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Optional[int] = ['longest', 'max_length', 'do_not_pad'] a__: List[Any] = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: Dict = feat_extract(lowercase , padding=lowercase , max_length=lowercase , return_tensors='np') a__: Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self.assertTrue(input_values[0][8_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self.assertTrue(input_values[0][10_00:].sum() < 1e-6) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Optional[int] = range(8_00 , 14_00 , 2_00) a__: List[str] = [floats_list((1, x))[0] for x in lengths] a__: Tuple = ['longest', 'max_length', 'do_not_pad'] a__: Dict = [None, 16_00, None] for max_length, padding in zip(lowercase , lowercase): a__: int = feat_extract(lowercase , max_length=lowercase , padding=lowercase) a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_00]) self._check_zero_mean_unit_variance(input_values[1][:10_00]) self._check_zero_mean_unit_variance(input_values[2][:12_00]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Any = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Dict = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='max_length' , return_tensors='np') a__: int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1]) self._check_zero_mean_unit_variance(input_values[2]) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: int = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: str = feat_extract( lowercase , truncation=lowercase , max_length=10_00 , padding='longest' , return_tensors='np') a__: Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 10_00)) a__: Dict = [floats_list((1, x))[0] for x in range(8_00 , 14_00 , 2_00)] a__: Tuple = feat_extract( lowercase , truncation=lowercase , max_length=20_00 , padding='longest' , return_tensors='np') a__: str = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_00]) self._check_zero_mean_unit_variance(input_values[1, :10_00]) self._check_zero_mean_unit_variance(input_values[2]) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 12_00)) @require_torch def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' import torch a__: Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) a__: Tuple = np.random.rand(1_00).astype(np.floataa) a__: Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a__: Any = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np') self.assertTrue(np_processed.input_values.dtype == np.floataa) a__: Optional[Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt') self.assertTrue(pt_processed.input_values.dtype == torch.floataa) @slow @require_torch def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: a__: str = WavaVecaConfig.from_pretrained(lowercase) a__: str = WavaVecaFeatureExtractor.from_pretrained(lowercase) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer')
290
1
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': { 'Salesforce/codegen-350M-mono': 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json', }, 'merges_file': { 'Salesforce/codegen-350M-mono': 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt', }, 'tokenizer_file': { 'Salesforce/codegen-350M-mono': ( 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json' ), }, } lowercase__ = { 'Salesforce/codegen-350M-mono': 2048, } class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = ["""input_ids""", """attention_mask"""] a__ = CodeGenTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase="<|endoftext|>" , lowercase="<|endoftext|>" , lowercase="<|endoftext|>" , lowercase=False , **lowercase , ) -> Optional[Any]: '''simple docstring''' super().__init__( lowercase , lowercase , tokenizer_file=lowercase , unk_token=lowercase , bos_token=lowercase , eos_token=lowercase , add_prefix_space=lowercase , **lowercase , ) if kwargs.pop('add_bos_token' , lowercase): a__: List[str] = kwargs.pop('name_or_path' , '') raise ValueError( 'Currenty GPT2\'s fast tokenizer does NOT support adding a BOS token.' 'Instead you should use GPT2\'s slow tokenizer class `CodeGenTokenizer` as follows: \n' f'`CodeGenTokenizer.from_pretrained(\'{model_id}\')`\nor\n' f'`AutoTokenizer.from_pretrained(\'{model_id}\', use_fast=False)`\n' 'This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.' ' so that the fast tokenizer works correctly.') a__: Any = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get('add_prefix_space' , lowercase) != add_prefix_space: a__: str = getattr(lowercase , pre_tok_state.pop('type')) a__: str = add_prefix_space a__: List[str] = pre_tok_class(**lowercase) a__: Dict = add_prefix_space def lowerCamelCase_ ( self , *lowercase , **lowercase) -> BatchEncoding: '''simple docstring''' a__: Optional[int] = kwargs.get('is_split_into_words' , lowercase) assert self.add_prefix_space or not is_split_into_words, ( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowercase , **lowercase) def lowerCamelCase_ ( self , *lowercase , **lowercase) -> BatchEncoding: '''simple docstring''' a__: Dict = kwargs.get('is_split_into_words' , lowercase) assert self.add_prefix_space or not is_split_into_words, ( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*lowercase , **lowercase) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' a__: Optional[Any] = self._tokenizer.model.save(lowercase , name=lowercase) return tuple(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase = False , lowercase = None , lowercase = None , **lowercase , ) -> str: '''simple docstring''' a__: int = super().decode( token_ids=lowercase , skip_special_tokens=lowercase , clean_up_tokenization_spaces=lowercase , **lowercase , ) if truncate_before_pattern is not None and len(lowercase) > 0: a__: str = self.truncate(lowercase , lowercase) return decoded_text def lowerCamelCase_ ( self , lowercase , lowercase) -> Dict: '''simple docstring''' def find_re(lowercase , lowercase , lowercase): a__: Tuple = pattern.search(lowercase , lowercase) return m.start() if m else -1 a__: List[str] = [re.compile(lowercase , re.MULTILINE) for pattern in truncate_before_pattern] a__: str = list(re.finditer('^print' , lowercase , re.MULTILINE)) if len(lowercase) > 1: a__: int = completion[: prints[1].start()] a__: Dict = list(re.finditer('^def' , lowercase , re.MULTILINE)) if len(lowercase) > 1: a__: str = completion[: defs[1].start()] a__: Optional[int] = 0 a__: int = [ pos for pos in [find_re(lowercase , lowercase , lowercase) for terminal in terminals] if pos != -1 ] if len(lowercase) > 0: return completion[: min(lowercase)] else: return completion
290
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class __snake_case ( __lowerCAmelCase ): a__ = """decision_transformer""" a__ = ["""past_key_values"""] a__ = { """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=17 , lowercase=4 , lowercase=1_28 , lowercase=40_96 , lowercase=True , lowercase=1 , lowercase=10_24 , lowercase=3 , lowercase=1 , lowercase=None , lowercase="relu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=1e-5 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=5_02_56 , lowercase=5_02_56 , lowercase=False , lowercase=False , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = state_dim a__: int = act_dim a__: List[Any] = hidden_size a__: List[str] = max_ep_len a__: List[Any] = action_tanh a__: Optional[Any] = vocab_size a__: Tuple = n_positions a__: Dict = n_layer a__: Optional[int] = n_head a__: Optional[int] = n_inner a__: Any = activation_function a__: Union[str, Any] = resid_pdrop a__: Any = embd_pdrop a__: Any = attn_pdrop a__: List[Any] = layer_norm_epsilon a__: Optional[Any] = initializer_range a__: Any = scale_attn_weights a__: Dict = use_cache a__: Optional[int] = scale_attn_by_inverse_layer_idx a__: List[str] = reorder_and_upcast_attn a__: Any = bos_token_id a__: int = eos_token_id super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase)
290
1
"""simple docstring""" from math import pow, sqrt def __a ( *_SCREAMING_SNAKE_CASE ) ->bool: a__: Union[str, Any] = len(_SCREAMING_SNAKE_CASE ) > 0 and all(value > 0.0 for value in values ) return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->float | ValueError: return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: while a != 0: a__ , a__: List[str] = b % a, a return b def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: if gcd(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) != 1: a__: Dict = F'mod inverse of {a!r} and {m!r} does not exist' raise ValueError(_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Union[str, Any] = 1, 0, a a__ , a__ , a__: Any = 0, 1, m while va != 0: a__: int = ua // va a__ , a__ , a__ , a__ , a__ , a__: Any = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
290
1
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class __snake_case ( __lowerCAmelCase ): def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: List[str] = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(lowercase , 'hidden_sizes')) self.parent.assertTrue(hasattr(lowercase , 'num_attention_heads')) self.parent.assertTrue(hasattr(lowercase , 'num_encoder_blocks')) class __snake_case : def __init__( self , lowercase , lowercase=13 , lowercase=64 , lowercase=3 , lowercase=4 , lowercase=[2, 2, 2, 2] , lowercase=[8, 4, 2, 1] , lowercase=[16, 32, 64, 1_28] , lowercase=[1, 4, 8, 16] , lowercase=[1, 2, 4, 8] , lowercase=True , lowercase=True , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.02 , lowercase=3 , lowercase=None , ) -> Any: '''simple docstring''' a__: Any = parent a__: List[Any] = batch_size a__: str = image_size a__: int = num_channels a__: List[Any] = num_encoder_blocks a__: Any = sr_ratios a__: Tuple = depths a__: Dict = hidden_sizes a__: Tuple = downsampling_rates a__: Union[str, Any] = num_attention_heads a__: str = is_training a__: List[Any] = use_labels a__: Dict = hidden_act a__: Dict = hidden_dropout_prob a__: Union[str, Any] = attention_probs_dropout_prob a__: Optional[Any] = initializer_range a__: Any = num_labels a__: Optional[Any] = scope def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) a__: Optional[Any] = None if self.use_labels: a__: str = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels) a__: Tuple = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self) -> str: '''simple docstring''' return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = SegformerModel(config=lowercase) model.to(lowercase) model.eval() a__: Tuple = model(lowercase) a__: str = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width)) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' a__: Optional[Any] = self.num_labels a__: Union[str, Any] = SegformerForSemanticSegmentation(lowercase) model.to(lowercase) model.eval() a__: int = model(lowercase) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4)) a__: int = model(lowercase , labels=lowercase) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4)) self.parent.assertGreater(result.loss , 0.0) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> str: '''simple docstring''' a__: Tuple = 1 a__: Optional[Any] = SegformerForSemanticSegmentation(config=lowercase) model.to(lowercase) model.eval() a__: List[str] = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size)).to(lowercase) a__: Union[str, Any] = model(lowercase , labels=lowercase) self.parent.assertGreater(result.loss , 0.0) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: str = self.prepare_config_and_inputs() a__ , a__ , a__: str = config_and_inputs a__: Optional[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) a__ = ( { """feature-extraction""": SegformerModel, """image-classification""": SegformerForImageClassification, """image-segmentation""": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) a__ = True a__ = False a__ = False a__ = False def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Optional[int] = SegformerModelTester(self) a__: Union[str, Any] = SegformerConfigTester(self , config_class=lowercase) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*lowercase) @unittest.skip('SegFormer does not use inputs_embeds') def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' pass @unittest.skip('SegFormer does not have get_input_embeddings method and get_output_embeddings methods') def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' pass def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__ , a__: Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__: Union[str, Any] = model_class(lowercase) a__: int = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__: Union[str, Any] = [*signature.parameters.keys()] a__: List[str] = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowercase) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__ , a__: Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() a__: List[Any] = True for model_class in self.all_model_classes: a__: Tuple = True a__: List[str] = False a__: Optional[Any] = True a__: Union[str, Any] = model_class(lowercase) model.to(lowercase) model.eval() with torch.no_grad(): a__: Dict = model(**self._prepare_for_class(lowercase , lowercase)) a__: Optional[Any] = outputs.attentions a__: Dict = sum(self.model_tester.depths) self.assertEqual(len(lowercase) , lowercase) # check that output_attentions also work using config del inputs_dict["output_attentions"] a__: List[str] = True a__: List[str] = model_class(lowercase) model.to(lowercase) model.eval() with torch.no_grad(): a__: str = model(**self._prepare_for_class(lowercase , lowercase)) a__: List[str] = outputs.attentions self.assertEqual(len(lowercase) , lowercase) # verify the first attentions (first block, first layer) a__: Optional[int] = (self.model_tester.image_size // 4) ** 2 a__: Any = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) a__: List[Any] = (self.model_tester.image_size // 32) ** 2 a__: Dict = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:]) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) a__: Tuple = len(lowercase) # Check attention is always last and order is fine a__: Union[str, Any] = True a__: Any = True a__: Optional[int] = model_class(lowercase) model.to(lowercase) model.eval() with torch.no_grad(): a__: int = model(**self._prepare_for_class(lowercase , lowercase)) self.assertEqual(out_len + 1 , len(lowercase)) a__: Dict = outputs.attentions self.assertEqual(len(lowercase) , lowercase) # verify the first attentions (first block, first layer) a__: Any = (self.model_tester.image_size // 4) ** 2 a__: Dict = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' def check_hidden_states_output(lowercase , lowercase , lowercase): a__: List[Any] = model_class(lowercase) model.to(lowercase) model.eval() with torch.no_grad(): a__: Optional[int] = model(**self._prepare_for_class(lowercase , lowercase)) a__: List[str] = outputs.hidden_states a__: Optional[int] = self.model_tester.num_encoder_blocks self.assertEqual(len(lowercase) , lowercase) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:]) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) a__ , a__: Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__: List[str] = True check_hidden_states_output(lowercase , lowercase , lowercase) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] a__: List[Any] = True check_hidden_states_output(lowercase , lowercase , lowercase) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' if not self.model_tester.is_training: return a__ , a__: Dict = self.model_tester.prepare_config_and_inputs_for_common() a__: Tuple = True for model_class in self.all_model_classes: if model_class in get_values(lowercase): continue a__: Union[str, Any] = model_class(lowercase) model.to(lowercase) model.train() a__: List[Any] = self._prepare_for_class(lowercase , lowercase , return_labels=lowercase) a__: List[str] = model(**lowercase).loss loss.backward() @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' pass @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__: Optional[Any] = SegformerModel.from_pretrained(lowercase) self.assertIsNotNone(lowercase) def __a ( ) ->Optional[Any]: a__: Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class __snake_case ( unittest.TestCase ): @slow def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: str = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=lowercase , align=lowercase , do_random_crop=lowercase) a__: List[str] = SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512').to( lowercase) a__: Optional[int] = prepare_img() a__: int = image_processor(images=lowercase , return_tensors='pt') a__: Dict = encoded_inputs.pixel_values.to(lowercase) with torch.no_grad(): a__: List[Any] = model(lowercase) a__: Tuple = torch.Size((1, model.config.num_labels, 1_28, 1_28)) self.assertEqual(outputs.logits.shape , lowercase) a__: Optional[Any] = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ]).to(lowercase) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , lowercase , atol=1e-4)) @slow def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: List[str] = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=lowercase , align=lowercase , do_random_crop=lowercase) a__: int = SegformerForSemanticSegmentation.from_pretrained( 'nvidia/segformer-b1-finetuned-cityscapes-1024-1024').to(lowercase) a__: str = prepare_img() a__: Optional[Any] = image_processor(images=lowercase , return_tensors='pt') a__: Dict = encoded_inputs.pixel_values.to(lowercase) with torch.no_grad(): a__: Optional[Any] = model(lowercase) a__: Tuple = torch.Size((1, model.config.num_labels, 1_28, 1_28)) self.assertEqual(outputs.logits.shape , lowercase) a__: Any = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ]).to(lowercase) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , lowercase , atol=1e-1)) @slow def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = SegformerImageProcessor( image_scale=(5_12, 5_12) , keep_ratio=lowercase , align=lowercase , do_random_crop=lowercase) a__: Optional[int] = SegformerForSemanticSegmentation.from_pretrained('nvidia/segformer-b0-finetuned-ade-512-512').to( lowercase) a__: Tuple = prepare_img() a__: Optional[Any] = image_processor(images=lowercase , return_tensors='pt') a__: Any = encoded_inputs.pixel_values.to(lowercase) with torch.no_grad(): a__: Tuple = model(lowercase) a__: Tuple = outputs.logits.detach().cpu() a__: Union[str, Any] = image_processor.post_process_semantic_segmentation(outputs=lowercase , target_sizes=[(5_00, 3_00)]) a__: str = torch.Size((5_00, 3_00)) self.assertEqual(segmentation[0].shape , lowercase) a__: Optional[Any] = image_processor.post_process_semantic_segmentation(outputs=lowercase) a__: List[str] = torch.Size((1_28, 1_28)) self.assertEqual(segmentation[0].shape , lowercase)
290
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex lowercase__ = logging.getLogger(__name__) class __snake_case : def __init__( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = False def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> str: '''simple docstring''' if not self.initialized: a__: Optional[int] = RagRetriever( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Optional[int] = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' self.retriever.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ , a__: str = self.retriever._main_retrieve(lowercase , lowercase) return doc_ids, retrieved_doc_embeds class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> int: '''simple docstring''' if index is not None and index.is_initialized() and len(lowercase) > 0: raise ValueError( 'When using Ray for distributed fine-tuning, ' 'you\'ll need to provide the paths instead, ' 'as the dataset and the index are loaded ' 'separately. More info in examples/rag/use_own_knowledge_dataset.py ') super().__init__( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , index=lowercase , init_retrieval=lowercase , ) a__: Any = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(lowercase , lowercase , lowercase , lowercase) for worker in self.retrieval_workers ]) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' logger.info('initializing retrieval') if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def lowerCamelCase_ ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if len(self.retrieval_workers) > 0: # Select a random retrieval actor. a__: int = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] a__ , a__: List[Any] = ray.get(random_worker.retrieve.remote(lowercase , lowercase)) else: a__ , a__: Dict = self._main_retrieve(lowercase , lowercase) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase=None , **lowercase) -> Tuple: '''simple docstring''' return super(lowercase , cls).get_tokenizers(lowercase , lowercase , **lowercase) @classmethod def lowerCamelCase_ ( cls , lowercase , lowercase , lowercase=None , **lowercase) -> Union[str, Any]: '''simple docstring''' a__: Optional[int] = kwargs.pop('config' , lowercase) or RagConfig.from_pretrained(lowercase , **lowercase) a__: Union[str, Any] = RagTokenizer.from_pretrained(lowercase , config=lowercase) a__: int = rag_tokenizer.question_encoder a__: Any = rag_tokenizer.generator if indexed_dataset is not None: a__: List[Any] = 'custom' a__: Optional[Any] = CustomHFIndex(config.retrieval_vector_size , lowercase) else: a__: Dict = cls._build_index(lowercase) return cls( lowercase , question_encoder_tokenizer=lowercase , generator_tokenizer=lowercase , retrieval_workers=lowercase , index=lowercase , )
290
1
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging lowercase__ = logging.get_logger(__name__) # TODO: upload to AWS lowercase__ = { 'yjernite/retribert-base-uncased': ( 'https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json' ), } class __snake_case ( __lowerCAmelCase ): a__ = """retribert""" def __init__( self , lowercase=3_05_22 , lowercase=7_68 , lowercase=8 , lowercase=12 , lowercase=30_72 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=2 , lowercase=0.02 , lowercase=1e-12 , lowercase=True , lowercase=1_28 , lowercase=0 , **lowercase , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=lowercase , **lowercase) a__: List[Any] = vocab_size a__: List[str] = hidden_size a__: int = num_hidden_layers a__: Dict = num_attention_heads a__: str = hidden_act a__: int = intermediate_size a__: Any = hidden_dropout_prob a__: List[str] = attention_probs_dropout_prob a__: List[Any] = max_position_embeddings a__: str = type_vocab_size a__: List[Any] = initializer_range a__: int = layer_norm_eps a__: Dict = share_encoders a__: Union[str, Any] = projection_dim
290
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: int = None if token is not None: a__: Tuple = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Optional[Any] = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: str = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) a__: int = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Dict = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_links.update({job['name']: job['html_url'] for job in result['jobs']} ) return job_links except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Dict: a__: Dict = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Dict = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: List[Any] = {} try: artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: Optional[int] = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() artifacts.update({artifact['name']: artifact['archive_download_url'] for artifact in result['artifacts']} ) return artifacts except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: List[Any] = None if token is not None: a__: Optional[int] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = result.headers['Location'] a__: Optional[int] = requests.get(_SCREAMING_SNAKE_CASE , allow_redirects=_SCREAMING_SNAKE_CASE ) a__: int = os.path.join(_SCREAMING_SNAKE_CASE , F'{artifact_name}.zip' ) with open(_SCREAMING_SNAKE_CASE , 'wb' ) as fp: fp.write(response.content ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: a__: List[Any] = [] a__: Optional[Any] = [] a__: List[Any] = None with zipfile.ZipFile(_SCREAMING_SNAKE_CASE ) as z: for filename in z.namelist(): if not os.path.isdir(_SCREAMING_SNAKE_CASE ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(_SCREAMING_SNAKE_CASE ) as f: for line in f: a__: Optional[int] = line.decode('UTF-8' ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs a__: Union[str, Any] = line[: line.index(': ' )] a__: Union[str, Any] = line[line.index(': ' ) + len(': ' ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith('FAILED ' ): # `test` is the test method that failed a__: Optional[int] = line[len('FAILED ' ) :] failed_tests.append(_SCREAMING_SNAKE_CASE ) elif filename == "job_name.txt": a__: Union[str, Any] = line if len(_SCREAMING_SNAKE_CASE ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( F'`errors` and `failed_tests` should have the same number of elements. Got {len(_SCREAMING_SNAKE_CASE )} for `errors` ' F'and {len(_SCREAMING_SNAKE_CASE )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some' ' problem.' ) a__: Tuple = None if job_name and job_links: a__: Dict = job_links.get(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # A list with elements of the form (line of error, error, failed test) a__: int = [x + [y] + [job_link] for x, y in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )] return result def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->str: a__: int = [] a__: Optional[int] = [os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for p in os.listdir(_SCREAMING_SNAKE_CASE ) if p.endswith('.zip' )] for p in paths: errors.extend(get_errors_from_single_artifact(_SCREAMING_SNAKE_CASE , job_links=_SCREAMING_SNAKE_CASE ) ) return errors def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Any: a__: str = Counter() counter.update([x[1] for x in logs] ) a__: int = counter.most_common() a__: Any = {} for error, count in counts: if error_filter is None or error not in error_filter: a__: List[str] = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} a__: Optional[Any] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: List[str] = test.split('::' )[0] if test.startswith('tests/models/' ): a__: Dict = test.split('/' )[2] else: a__: Any = None return test def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: int = [(x[0], x[1], get_model(x[2] )) for x in logs] a__: List[Any] = [x for x in logs if x[2] is not None] a__: Optional[Any] = {x[2] for x in logs} a__: Dict = {} for test in tests: a__: Union[str, Any] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) a__: Union[str, Any] = counter.most_common() a__: List[str] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} a__: List[Any] = sum(error_counts.values() ) if n_errors > 0: a__: Any = {'count': n_errors, 'errors': error_counts} a__: Optional[int] = dict(sorted(r.items() , key=lambda _SCREAMING_SNAKE_CASE : item[1]["count"] , reverse=_SCREAMING_SNAKE_CASE ) ) return r def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Any = '| no. | error | status |' a__: Any = '|-:|:-|:-|' a__: str = [header, sep] for error in reduced_by_error: a__: int = reduced_by_error[error]['count'] a__: Tuple = F'| {count} | {error[:100]} | |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE ) ->str: a__: List[str] = '| model | no. of errors | major error | count |' a__: str = '|-:|-:|-:|-:|' a__: int = [header, sep] for model in reduced_by_model: a__: Tuple = reduced_by_model[model]['count'] a__ , a__: Dict = list(reduced_by_model[model]['errors'].items() )[0] a__: Dict = F'| {model} | {count} | {error[:60]} | {_count} |' lines.append(_SCREAMING_SNAKE_CASE ) return "\n".join(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') lowercase__ = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) lowercase__ = get_job_links(args.workflow_run_id, token=args.token) lowercase__ = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: lowercase__ = k.find(' / ') lowercase__ = k[index + len(' / ') :] lowercase__ = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) lowercase__ = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) lowercase__ = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error lowercase__ = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors lowercase__ = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) lowercase__ = reduce_by_error(errors) lowercase__ = reduce_by_model(errors) lowercase__ = make_github_table(reduced_by_error) lowercase__ = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
290
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'google/mobilenet_v1_1.0_224': 'https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json', 'google/mobilenet_v1_0.75_192': 'https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class __snake_case ( __lowerCAmelCase ): a__ = """mobilenet_v1""" def __init__( self , lowercase=3 , lowercase=2_24 , lowercase=1.0 , lowercase=8 , lowercase="relu6" , lowercase=True , lowercase=0.999 , lowercase=0.02 , lowercase=0.001 , **lowercase , ) -> Any: '''simple docstring''' super().__init__(**lowercase) if depth_multiplier <= 0: raise ValueError('depth_multiplier must be greater than zero.') a__: str = num_channels a__: List[str] = image_size a__: Dict = depth_multiplier a__: Optional[Any] = min_depth a__: Any = hidden_act a__: Tuple = tf_padding a__: List[Any] = classifier_dropout_prob a__: Any = initializer_range a__: Dict = layer_norm_eps class __snake_case ( __lowerCAmelCase ): a__ = version.parse("""1.11""" ) @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict([('pixel_values', {0: 'batch'})]) @property def lowerCamelCase_ ( self) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "image-classification": return OrderedDict([('logits', {0: 'batch'})]) else: return OrderedDict([('last_hidden_state', {0: 'batch'}), ('pooler_output', {0: 'batch'})]) @property def lowerCamelCase_ ( self) -> float: '''simple docstring''' return 1e-4
290
"""simple docstring""" import math def __a ( _SCREAMING_SNAKE_CASE ) ->bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_SCREAMING_SNAKE_CASE ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __a ( _SCREAMING_SNAKE_CASE = 0.1 ) ->int: a__: str = 3 a__: Optional[Any] = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_SCREAMING_SNAKE_CASE ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_electra import ElectraTokenizer lowercase__ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': { 'google/electra-small-generator': ( 'https://huggingface.co/google/electra-small-generator/resolve/main/vocab.txt' ), 'google/electra-base-generator': 'https://huggingface.co/google/electra-base-generator/resolve/main/vocab.txt', 'google/electra-large-generator': ( 'https://huggingface.co/google/electra-large-generator/resolve/main/vocab.txt' ), 'google/electra-small-discriminator': ( 'https://huggingface.co/google/electra-small-discriminator/resolve/main/vocab.txt' ), 'google/electra-base-discriminator': ( 'https://huggingface.co/google/electra-base-discriminator/resolve/main/vocab.txt' ), 'google/electra-large-discriminator': ( 'https://huggingface.co/google/electra-large-discriminator/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'google/electra-small-generator': ( 'https://huggingface.co/google/electra-small-generator/resolve/main/tokenizer.json' ), 'google/electra-base-generator': ( 'https://huggingface.co/google/electra-base-generator/resolve/main/tokenizer.json' ), 'google/electra-large-generator': ( 'https://huggingface.co/google/electra-large-generator/resolve/main/tokenizer.json' ), 'google/electra-small-discriminator': ( 'https://huggingface.co/google/electra-small-discriminator/resolve/main/tokenizer.json' ), 'google/electra-base-discriminator': ( 'https://huggingface.co/google/electra-base-discriminator/resolve/main/tokenizer.json' ), 'google/electra-large-discriminator': ( 'https://huggingface.co/google/electra-large-discriminator/resolve/main/tokenizer.json' ), }, } lowercase__ = { 'google/electra-small-generator': 512, 'google/electra-base-generator': 512, 'google/electra-large-generator': 512, 'google/electra-small-discriminator': 512, 'google/electra-base-discriminator': 512, 'google/electra-large-discriminator': 512, } lowercase__ = { 'google/electra-small-generator': {'do_lower_case': True}, 'google/electra-base-generator': {'do_lower_case': True}, 'google/electra-large-generator': {'do_lower_case': True}, 'google/electra-small-discriminator': {'do_lower_case': True}, 'google/electra-base-discriminator': {'do_lower_case': True}, 'google/electra-large-discriminator': {'do_lower_case': True}, } class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_INIT_CONFIGURATION a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = ElectraTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase="[UNK]" , lowercase="[SEP]" , lowercase="[PAD]" , lowercase="[CLS]" , lowercase="[MASK]" , lowercase=True , lowercase=None , **lowercase , ) -> List[Any]: '''simple docstring''' super().__init__( lowercase , tokenizer_file=lowercase , do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , tokenize_chinese_chars=lowercase , strip_accents=lowercase , **lowercase , ) a__: Any = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('lowercase' , lowercase) != do_lower_case or normalizer_state.get('strip_accents' , lowercase) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowercase) != tokenize_chinese_chars ): a__: Optional[Any] = getattr(lowercase , normalizer_state.pop('type')) a__: str = do_lower_case a__: Union[str, Any] = strip_accents a__: Union[str, Any] = tokenize_chinese_chars a__: Any = normalizer_class(**lowercase) a__: List[str] = do_lower_case def lowerCamelCase_ ( self , lowercase , lowercase=None) -> int: '''simple docstring''' a__: List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' a__: List[Any] = [self.sep_token_id] a__: Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def lowerCamelCase_ ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' a__: List[Any] = self._tokenizer.model.save(lowercase , name=lowercase) return tuple(lowercase)
290
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ = {'processing_wav2vec2_with_lm': ['Wav2Vec2ProcessorWithLM']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ = { 'configuration_deberta': ['DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DebertaConfig', 'DebertaOnnxConfig'], 'tokenization_deberta': ['DebertaTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ['DebertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'DebertaForMaskedLM', 'DebertaForQuestionAnswering', 'DebertaForSequenceClassification', 'DebertaForTokenClassification', 'DebertaModel', 'DebertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ 'TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDebertaForMaskedLM', 'TFDebertaForQuestionAnswering', 'TFDebertaForSequenceClassification', 'TFDebertaForTokenClassification', 'TFDebertaModel', 'TFDebertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
290
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = KandinskyInpaintPipeline a__ = ["""prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] a__ = [ """prompt""", """negative_prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] a__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """negative_prompt""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] a__ = False @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' return 32 @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim @property def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return 1_00 @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[int] = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base') return tokenizer @property def lowerCamelCase_ ( self) -> Any: '''simple docstring''' torch.manual_seed(0) a__: Dict = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=10_05 , ) a__: Optional[Any] = MultilingualCLIP(lowercase) a__: int = text_encoder.eval() return text_encoder @property def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' torch.manual_seed(0) a__: Any = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } a__: str = UNetaDConditionModel(**lowercase) return model @property def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' torch.manual_seed(0) a__: Any = VQModel(**self.dummy_movq_kwargs) return model def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Dict = self.dummy_text_encoder a__: int = self.dummy_tokenizer a__: str = self.dummy_unet a__: Any = self.dummy_movq a__: Tuple = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.00085 , beta_end=0.012 , clip_sample=lowercase , set_alpha_to_one=lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=lowercase , ) a__: Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowerCamelCase_ ( self , lowercase , lowercase=0) -> Any: '''simple docstring''' a__: List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase)).to(lowercase) a__: int = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1)).to(lowercase) # create init_image a__: Optional[int] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) a__: int = image.cpu().permute(0 , 2 , 3 , 1)[0] a__: Optional[int] = Image.fromarray(np.uinta(lowercase)).convert('RGB').resize((2_56, 2_56)) # create mask a__: Tuple = np.ones((64, 64) , dtype=np.floataa) a__: Optional[Any] = 0 if str(lowercase).startswith('mps'): a__: str = torch.manual_seed(lowercase) else: a__: Dict = torch.Generator(device=lowercase).manual_seed(lowercase) a__: Optional[int] = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Optional[Any] = 'cpu' a__: List[Any] = self.get_dummy_components() a__: Optional[Any] = self.pipeline_class(**lowercase) a__: str = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__: Optional[int] = pipe(**self.get_dummy_inputs(lowercase)) a__: List[str] = output.images a__: int = pipe( **self.get_dummy_inputs(lowercase) , return_dict=lowercase , )[0] a__: Optional[Any] = image[0, -3:, -3:, -1] a__: List[Any] = image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}') assert image.shape == (1, 64, 64, 3) a__: str = np.array( [0.8326919, 0.73790467, 0.20918581, 0.9309612, 0.5511791, 0.43713328, 0.5513321, 0.49922934, 0.59497786]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def lowerCamelCase_ ( self) -> str: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy') a__: int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png') a__: Union[str, Any] = np.ones((7_68, 7_68) , dtype=np.floataa) a__: int = 0 a__: Optional[int] = 'a hat' a__: int = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa) pipe_prior.to(lowercase) a__: Any = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa) a__: Optional[Any] = pipeline.to(lowercase) pipeline.set_progress_bar_config(disable=lowercase) a__: Dict = torch.Generator(device='cpu').manual_seed(0) a__ , a__: Optional[Any] = pipe_prior( lowercase , generator=lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() a__: List[str] = pipeline( lowercase , image=lowercase , mask_image=lowercase , image_embeds=lowercase , negative_image_embeds=lowercase , generator=lowercase , num_inference_steps=1_00 , height=7_68 , width=7_68 , output_type='np' , ) a__: str = output.images[0] assert image.shape == (7_68, 7_68, 3) assert_mean_pixel_difference(lowercase , lowercase)
290
1
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def __a ( ) ->tuple[list[int], int]: a__: Optional[int] = [randint(-1000 , 1000 ) for i in range(10 )] a__: int = randint(-5000 , 5000 ) return (arr, r) lowercase__ = make_dataset() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->tuple[int, ...]: for triplet in permutations(_SCREAMING_SNAKE_CASE , 3 ): if sum(_SCREAMING_SNAKE_CASE ) == target: return tuple(sorted(_SCREAMING_SNAKE_CASE ) ) return (0, 0, 0) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->tuple[int, int, int]: arr.sort() a__: Dict = len(_SCREAMING_SNAKE_CASE ) for i in range(n - 1 ): a__ , a__: int = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def __a ( ) ->tuple[float, float]: a__: Optional[int] = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' a__: int = '\ntriplet_sum1(*dataset)\n' a__: List[str] = '\ntriplet_sum2(*dataset)\n' a__: List[Any] = repeat(setup=_SCREAMING_SNAKE_CASE , stmt=_SCREAMING_SNAKE_CASE , repeat=5 , number=10000 ) a__: Tuple = repeat(setup=_SCREAMING_SNAKE_CASE , stmt=_SCREAMING_SNAKE_CASE , repeat=5 , number=10000 ) return (min(_SCREAMING_SNAKE_CASE ), min(_SCREAMING_SNAKE_CASE )) if __name__ == "__main__": from doctest import testmod testmod() lowercase__ = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
290
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() lowercase__ = logging.get_logger('transformers.models.encodec') lowercase__ = { 'quantizer.vq.layers.*._codebook.inited': 'quantizer.layers.*.codebook.inited', 'quantizer.vq.layers.*._codebook.cluster_size': 'quantizer.layers.*.codebook.cluster_size', 'quantizer.vq.layers.*._codebook.embed': 'quantizer.layers.*.codebook.embed', 'quantizer.vq.layers.*._codebook.embed_avg': 'quantizer.layers.*.codebook.embed_avg', } lowercase__ = { 'encoder.model.0.conv.conv': 'encoder.layers.0.conv', 'encoder.model.1.block.1.conv.conv': 'encoder.layers.1.block.1.conv', 'encoder.model.1.block.3.conv.conv': 'encoder.layers.1.block.3.conv', 'encoder.model.1.shortcut.conv.conv': 'encoder.layers.1.shortcut.conv', 'encoder.model.3.conv.conv': 'encoder.layers.3.conv', 'encoder.model.4.block.1.conv.conv': 'encoder.layers.4.block.1.conv', 'encoder.model.4.block.3.conv.conv': 'encoder.layers.4.block.3.conv', 'encoder.model.4.shortcut.conv.conv': 'encoder.layers.4.shortcut.conv', 'encoder.model.6.conv.conv': 'encoder.layers.6.conv', 'encoder.model.7.block.1.conv.conv': 'encoder.layers.7.block.1.conv', 'encoder.model.7.block.3.conv.conv': 'encoder.layers.7.block.3.conv', 'encoder.model.7.shortcut.conv.conv': 'encoder.layers.7.shortcut.conv', 'encoder.model.9.conv.conv': 'encoder.layers.9.conv', 'encoder.model.10.block.1.conv.conv': 'encoder.layers.10.block.1.conv', 'encoder.model.10.block.3.conv.conv': 'encoder.layers.10.block.3.conv', 'encoder.model.10.shortcut.conv.conv': 'encoder.layers.10.shortcut.conv', 'encoder.model.12.conv.conv': 'encoder.layers.12.conv', 'encoder.model.13.lstm': 'encoder.layers.13.lstm', 'encoder.model.15.conv.conv': 'encoder.layers.15.conv', } lowercase__ = { 'encoder.model.0.conv.norm': 'encoder.layers.0.norm', 'encoder.model.1.block.1.conv.norm': 'encoder.layers.1.block.1.norm', 'encoder.model.1.block.3.conv.norm': 'encoder.layers.1.block.3.norm', 'encoder.model.1.shortcut.conv.norm': 'encoder.layers.1.shortcut.norm', 'encoder.model.3.conv.norm': 'encoder.layers.3.norm', 'encoder.model.4.block.1.conv.norm': 'encoder.layers.4.block.1.norm', 'encoder.model.4.block.3.conv.norm': 'encoder.layers.4.block.3.norm', 'encoder.model.4.shortcut.conv.norm': 'encoder.layers.4.shortcut.norm', 'encoder.model.6.conv.norm': 'encoder.layers.6.norm', 'encoder.model.7.block.1.conv.norm': 'encoder.layers.7.block.1.norm', 'encoder.model.7.block.3.conv.norm': 'encoder.layers.7.block.3.norm', 'encoder.model.7.shortcut.conv.norm': 'encoder.layers.7.shortcut.norm', 'encoder.model.9.conv.norm': 'encoder.layers.9.norm', 'encoder.model.10.block.1.conv.norm': 'encoder.layers.10.block.1.norm', 'encoder.model.10.block.3.conv.norm': 'encoder.layers.10.block.3.norm', 'encoder.model.10.shortcut.conv.norm': 'encoder.layers.10.shortcut.norm', 'encoder.model.12.conv.norm': 'encoder.layers.12.norm', 'encoder.model.15.conv.norm': 'encoder.layers.15.norm', } lowercase__ = { 'decoder.model.0.conv.conv': 'decoder.layers.0.conv', 'decoder.model.1.lstm': 'decoder.layers.1.lstm', 'decoder.model.3.convtr.convtr': 'decoder.layers.3.conv', 'decoder.model.4.block.1.conv.conv': 'decoder.layers.4.block.1.conv', 'decoder.model.4.block.3.conv.conv': 'decoder.layers.4.block.3.conv', 'decoder.model.4.shortcut.conv.conv': 'decoder.layers.4.shortcut.conv', 'decoder.model.6.convtr.convtr': 'decoder.layers.6.conv', 'decoder.model.7.block.1.conv.conv': 'decoder.layers.7.block.1.conv', 'decoder.model.7.block.3.conv.conv': 'decoder.layers.7.block.3.conv', 'decoder.model.7.shortcut.conv.conv': 'decoder.layers.7.shortcut.conv', 'decoder.model.9.convtr.convtr': 'decoder.layers.9.conv', 'decoder.model.10.block.1.conv.conv': 'decoder.layers.10.block.1.conv', 'decoder.model.10.block.3.conv.conv': 'decoder.layers.10.block.3.conv', 'decoder.model.10.shortcut.conv.conv': 'decoder.layers.10.shortcut.conv', 'decoder.model.12.convtr.convtr': 'decoder.layers.12.conv', 'decoder.model.13.block.1.conv.conv': 'decoder.layers.13.block.1.conv', 'decoder.model.13.block.3.conv.conv': 'decoder.layers.13.block.3.conv', 'decoder.model.13.shortcut.conv.conv': 'decoder.layers.13.shortcut.conv', 'decoder.model.15.conv.conv': 'decoder.layers.15.conv', } lowercase__ = { 'decoder.model.0.conv.norm': 'decoder.layers.0.norm', 'decoder.model.3.convtr.norm': 'decoder.layers.3.norm', 'decoder.model.4.block.1.conv.norm': 'decoder.layers.4.block.1.norm', 'decoder.model.4.block.3.conv.norm': 'decoder.layers.4.block.3.norm', 'decoder.model.4.shortcut.conv.norm': 'decoder.layers.4.shortcut.norm', 'decoder.model.6.convtr.norm': 'decoder.layers.6.norm', 'decoder.model.7.block.1.conv.norm': 'decoder.layers.7.block.1.norm', 'decoder.model.7.block.3.conv.norm': 'decoder.layers.7.block.3.norm', 'decoder.model.7.shortcut.conv.norm': 'decoder.layers.7.shortcut.norm', 'decoder.model.9.convtr.norm': 'decoder.layers.9.norm', 'decoder.model.10.block.1.conv.norm': 'decoder.layers.10.block.1.norm', 'decoder.model.10.block.3.conv.norm': 'decoder.layers.10.block.3.norm', 'decoder.model.10.shortcut.conv.norm': 'decoder.layers.10.shortcut.norm', 'decoder.model.12.convtr.norm': 'decoder.layers.12.norm', 'decoder.model.13.block.1.conv.norm': 'decoder.layers.13.block.1.norm', 'decoder.model.13.block.3.conv.norm': 'decoder.layers.13.block.3.norm', 'decoder.model.13.shortcut.conv.norm': 'decoder.layers.13.shortcut.norm', 'decoder.model.15.conv.norm': 'decoder.layers.15.norm', } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } lowercase__ = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } lowercase__ = [] lowercase__ = [] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: for attribute in key.split('.' ): a__: str = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: a__: List[str] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: a__: Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": a__: str = value elif weight_type == "weight_g": a__: int = value elif weight_type == "weight_v": a__: Tuple = value elif weight_type == "bias": a__: Dict = value elif weight_type == "running_mean": a__: Any = value elif weight_type == "running_var": a__: Tuple = value elif weight_type == "num_batches_tracked": a__: List[str] = value elif weight_type == "weight_ih_l0": a__: List[Any] = value elif weight_type == "weight_hh_l0": a__: List[Any] = value elif weight_type == "bias_ih_l0": a__: List[Any] = value elif weight_type == "bias_hh_l0": a__: List[Any] = value elif weight_type == "weight_ih_l1": a__: int = value elif weight_type == "weight_hh_l1": a__: str = value elif weight_type == "bias_ih_l1": a__: Union[str, Any] = value elif weight_type == "bias_hh_l1": a__: Any = value else: a__: Union[str, Any] = value logger.info(F'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: for key in ignore_keys: if key.endswith('.*' ): if name.startswith(key[:-1] ): return True elif ".*." in key: a__ , a__: Optional[Any] = key.split('.*.' ) if prefix in name and suffix in name: return True elif key in name: return True return False def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: List[Any] = [] if model_name == "encodec_24khz" or "encodec_32khz": a__: Optional[int] = MAPPING_24K elif model_name == "encodec_48khz": a__: List[Any] = MAPPING_48K else: raise ValueError(F'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): logger.info(F'{name} was ignored' ) continue a__: int = False for key, mapped_key in MAPPING.items(): if "*" in key: a__ , a__: str = key.split('.*.' ) if prefix in name and suffix in name: a__: List[str] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('embed' ) and name.endswith('embed_avg' ): continue a__: List[str] = True if "*" in mapped_key: a__: List[str] = name.split(_SCREAMING_SNAKE_CASE )[0].split('.' )[-2] a__: str = mapped_key.replace('*' , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: a__: int = 'weight_g' elif "weight_v" in name: a__: Dict = 'weight_v' elif "weight_ih_l0" in name: a__: int = 'weight_ih_l0' elif "weight_hh_l0" in name: a__: Union[str, Any] = 'weight_hh_l0' elif "bias_ih_l0" in name: a__: Optional[Any] = 'bias_ih_l0' elif "bias_hh_l0" in name: a__: Optional[int] = 'bias_hh_l0' elif "weight_ih_l1" in name: a__: Dict = 'weight_ih_l1' elif "weight_hh_l1" in name: a__: Optional[Any] = 'weight_hh_l1' elif "bias_ih_l1" in name: a__: List[str] = 'bias_ih_l1' elif "bias_hh_l1" in name: a__: Optional[Any] = 'bias_hh_l1' elif "bias" in name: a__: List[str] = 'bias' elif "weight" in name: a__: Any = 'weight' elif "running_mean" in name: a__: Dict = 'running_mean' elif "running_var" in name: a__: Dict = 'running_var' elif "num_batches_tracked" in name: a__: Dict = 'num_batches_tracked' else: a__: List[str] = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(F'Unused weights: {unused_weights}' ) @torch.no_grad() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) ->int: if config_path is not None: a__: Dict = EncodecConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: a__: Tuple = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": a__: Any = [8, 5, 4, 4] a__: List[str] = [2.2] a__: List[Any] = 64 a__: Dict = 32000 a__: Union[str, Any] = 2048 a__: Union[str, Any] = False a__: Any = False a__: Optional[Any] = False elif model_name == "encodec_48khz": a__: Optional[int] = [8, 5, 4, 2] a__: Union[str, Any] = [3.0, 6.0, 12.0, 24.0] a__: List[str] = 48000 a__: Tuple = 2 a__: Optional[Any] = False a__: Optional[int] = 'time_group_norm' a__: Union[str, Any] = True a__: Dict = 1.0 a__: str = 0.01 else: raise ValueError(F'Unknown model name: {model_name}' ) a__: Optional[int] = EncodecModel(_SCREAMING_SNAKE_CASE ) a__: List[str] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(_SCREAMING_SNAKE_CASE ) a__: int = torch.load(_SCREAMING_SNAKE_CASE ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights a__: str = original_checkpoint['best_state'] recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) if repo_id: print('Pushing to the hub...' ) feature_extractor.push_to_hub(_SCREAMING_SNAKE_CASE ) model.push_to_hub(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) lowercase__ = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
290
1
"""simple docstring""" import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class __snake_case ( unittest.TestCase , __lowerCAmelCase ): def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: Dict = load_tool('text-to-speech') self.tool.setup() def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' torch.manual_seed(0) a__: Optional[Any] = self.tool('hey') a__: Optional[Any] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485]) , )) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' torch.manual_seed(0) a__: Union[str, Any] = self.tool('hey') a__: Union[str, Any] = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485]) , ))
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: if height >= 1: move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_disk(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) move_tower(height - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: print('moving disk from' , _SCREAMING_SNAKE_CASE , 'to' , _SCREAMING_SNAKE_CASE ) def __a ( ) ->List[str]: a__: Dict = int(input('Height of hanoi: ' ).strip() ) move_tower(_SCREAMING_SNAKE_CASE , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
290
1
"""simple docstring""" import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def __a ( _SCREAMING_SNAKE_CASE ) ->Dict: # picklable for multiprocessing return x.sum() def __a ( _SCREAMING_SNAKE_CASE ) ->Dict: # picklable for multiprocessing return i + 1 @dataclass class __snake_case : a__ = 42 a__ = 42 class __snake_case ( __lowerCAmelCase ): def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: List[str] = {} a__: Optional[int] = [] a__: List[Any] = 1 a__: str = [1, 2] a__: Tuple = {'a': 1, 'b': 2} a__: Dict = {'a': [1, 2], 'b': [3, 4]} a__: List[str] = {'a': {'1': 1}, 'b': 2} a__: Any = {'a': 1, 'b': 2, 'c': 3, 'd': 4} a__: int = {} a__: Union[str, Any] = [] a__: Optional[int] = 2 a__: Optional[int] = [2, 3] a__: List[Any] = {'a': 2, 'b': 3} a__: Optional[Any] = {'a': [2, 3], 'b': [4, 5]} a__: List[str] = {'a': {'1': 2}, 'b': 3} a__: Tuple = {'a': 2, 'b': 3, 'c': 4, 'd': 5} self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase) , lowercase) a__: Tuple = 2 self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) self.assertEqual(map_nested(lowercase , lowercase , num_proc=lowercase) , lowercase) a__: List[Any] = {'a': np.eye(2), 'b': np.zeros(3), 'c': np.ones(2)} a__: Dict = {'a': 2, 'b': 0, 'c': 2} a__: Dict = { 'a': np.eye(2).astype(lowercase), 'b': np.zeros(3).astype(lowercase), 'c': np.ones(2).astype(lowercase), } self.assertEqual(map_nested(lowercase , lowercase , map_numpy=lowercase) , lowercase) self.assertEqual( {k: v.tolist() for k, v in map_nested(lowercase , lowercase , map_numpy=lowercase).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(lowercase , lowercase , map_numpy=lowercase , num_proc=lowercase) , lowercase) self.assertEqual( {k: v.tolist() for k, v in map_nested(lowercase , lowercase , map_numpy=lowercase , num_proc=lowercase).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(lowercase): # can't pickle a local lambda map_nested(lambda lowercase: x + 1 , lowercase , num_proc=lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = {'a': 1, 'b': 2} a__: Tuple = {'a': 3, 'b': 4} a__: Optional[int] = {'a': 5, 'b': 6} a__: Any = sorted([('a', (1, 3, 5)), ('b', (2, 4, 6))]) self.assertEqual(sorted(zip_dict(lowercase , lowercase , lowercase)) , lowercase) def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' class __snake_case : a__ = """bar""" a__: Optional[Any] = Foo() self.assertEqual(foo.my_attr , 'bar') with temporary_assignment(lowercase , 'my_attr' , 'BAR'): self.assertEqual(foo.my_attr , 'BAR') self.assertEqual(foo.my_attr , 'bar') @pytest.mark.parametrize( 'iterable_length, num_proc, expected_num_proc' , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: with patch('datasets.utils.py_utils._single_map_nested' ) as mock_single_map_nested, patch( 'datasets.parallel.parallel.Pool' ) as mock_multiprocessing_pool: a__: int = {F'{i}': i for i in range(_SCREAMING_SNAKE_CASE )} a__: List[str] = map_nested(lambda _SCREAMING_SNAKE_CASE : x + 10 , _SCREAMING_SNAKE_CASE , num_proc=_SCREAMING_SNAKE_CASE , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class __snake_case ( __lowerCAmelCase ): @require_tf def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers a__: str = layers.Dense(2) def gen_random_output(): a__: str = tf.random.uniform((1, 3)) return model(lowercase).numpy() with temp_seed(42 , set_tensorflow=lowercase): a__: List[str] = gen_random_output() with temp_seed(42 , set_tensorflow=lowercase): a__: Dict = gen_random_output() a__: List[Any] = gen_random_output() np.testing.assert_equal(lowercase , lowercase) self.assertGreater(np.abs(outa - outa).sum() , 0) @require_torch def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' import torch def gen_random_output(): a__: Optional[int] = torch.nn.Linear(3 , 2) a__: str = torch.rand(1 , 3) return model(lowercase).detach().numpy() with temp_seed(42 , set_pytorch=lowercase): a__: int = gen_random_output() with temp_seed(42 , set_pytorch=lowercase): a__: Optional[int] = gen_random_output() a__: int = gen_random_output() np.testing.assert_equal(lowercase , lowercase) self.assertGreater(np.abs(outa - outa).sum() , 0) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3) with temp_seed(42): a__: List[Any] = gen_random_output() with temp_seed(42): a__: Optional[int] = gen_random_output() a__: Optional[int] = gen_random_output() np.testing.assert_equal(lowercase , lowercase) self.assertGreater(np.abs(outa - outa).sum() , 0) @pytest.mark.parametrize('input_data' , [{}] ) def __a ( _SCREAMING_SNAKE_CASE ) ->List[Any]: a__: Any = NestedDataStructure(_SCREAMING_SNAKE_CASE ).data assert output_data == input_data @pytest.mark.parametrize( 'data, expected_output' , [ ({}, []), ([], []), ('foo', ['foo']), (['foo', 'bar'], ['foo', 'bar']), ([['foo', 'bar']], ['foo', 'bar']), ([[['foo'], ['bar']]], ['foo', 'bar']), ([[['foo'], 'bar']], ['foo', 'bar']), ({'a': 1, 'b': 2}, [1, 2]), ({'a': [1, 2], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[1, 2]], 'b': [[3, 4]]}, [1, 2, 3, 4]), ({'a': [[1, 2]], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [[[3], [4]]]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [[3, 4]]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [3, 4]}, [1, 2, 3, 4]), ({'a': [[[1], [2]]], 'b': [3, [4]]}, [1, 2, 3, 4]), ({'a': {'1': 1}, 'b': 2}, [1, 2]), ({'a': {'1': [1]}, 'b': 2}, [1, 2]), ({'a': {'1': [1]}, 'b': [2]}, [1, 2]), ] , ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: a__: List[Any] = NestedDataStructure(_SCREAMING_SNAKE_CASE ).flatten() assert output == expected_output def __a ( ) ->Dict: a__: Dict = A(x=1 , y='foobar' ) a__: Optional[Any] = {'x': 1, 'y': 'foobar'} assert asdict(_SCREAMING_SNAKE_CASE ) == expected_output a__: List[Any] = {'a': {'b': A(x=10 , y='foo' )}, 'c': [A(x=20 , y='bar' )]} a__: int = {'a': {'b': {'x': 10, 'y': 'foo'}}, 'c': [{'x': 20, 'y': 'bar'}]} assert asdict(_SCREAMING_SNAKE_CASE ) == expected_output with pytest.raises(_SCREAMING_SNAKE_CASE ): asdict([1, A(x=10 , y='foo' )] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Tuple: return text.split() def __a ( _SCREAMING_SNAKE_CASE ) ->Dict: yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def __a ( ) ->Dict: with Pool(2 ) as pool: a__: str = list(iflatmap_unordered(_SCREAMING_SNAKE_CASE , _split_text , kwargs_iterable=[{'text': 'hello there'}] * 10 ) ) assert out.count('hello' ) == 10 assert out.count('there' ) == 10 assert len(_SCREAMING_SNAKE_CASE ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: a__: Any = list(iflatmap_unordered(_SCREAMING_SNAKE_CASE , _split_text , kwargs_iterable=[{'text': 'hello there'}] * 10 ) ) assert out.count('hello' ) == 10 assert out.count('there' ) == 10 assert len(_SCREAMING_SNAKE_CASE ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: a__: Tuple = [] for yield_time, content in iflatmap_unordered( _SCREAMING_SNAKE_CASE , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{'content': 'a'}, {'content': 'b'}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_SCREAMING_SNAKE_CASE ) assert out.count('a' ) == 2 assert out.count('b' ) == 2 assert len(_SCREAMING_SNAKE_CASE ) == 4
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = False ) ->str: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = F'Expected string as input, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[str] = F'Expected boolean as use_pascal parameter, found {type(_SCREAMING_SNAKE_CASE )}' raise ValueError(_SCREAMING_SNAKE_CASE ) a__: int = input_str.split('_' ) a__: List[str] = 0 if use_pascal else 1 a__: List[str] = words[start_index:] a__: List[str] = [word[0].upper() + word[1:] for word in words_to_capitalize] a__: List[str] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: a__: Dict = mf_knapsack(i - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: a__: List[str] = max( mf_knapsack(i - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , mf_knapsack(i - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , j - wt[i - 1] ) + val[i - 1] , ) a__: Union[str, Any] = val return f[i][j] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: str = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: a__: int = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: a__: Union[str, Any] = dp[i - 1][w_] return dp[n][w_], dp def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: if not (isinstance(_SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(_SCREAMING_SNAKE_CASE , (list, tuple) )): raise ValueError( 'Both the weights and values vectors must be either lists or tuples' ) a__: int = len(_SCREAMING_SNAKE_CASE ) if num_items != len(_SCREAMING_SNAKE_CASE ): a__: Optional[int] = ( 'The number of weights must be the same as the number of values.\n' F'But got {num_items} weights and {len(_SCREAMING_SNAKE_CASE )} values' ) raise ValueError(_SCREAMING_SNAKE_CASE ) for i in range(_SCREAMING_SNAKE_CASE ): if not isinstance(wt[i] , _SCREAMING_SNAKE_CASE ): a__: List[Any] = ( 'All weights must be integers but got weight of ' F'type {type(wt[i] )} at index {i}' ) raise TypeError(_SCREAMING_SNAKE_CASE ) a__ , a__: List[Any] = knapsack(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: set = set() _construct_solution(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return optimal_val, example_optional_set def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: # for the current item i at a maximum weight j to be part of an optimal subset, # the optimal value at (i, j) must be greater than the optimal value at (i-1, j). # where i - 1 means considering only the previous items at the given maximum weight if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , i - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: optimal_set.add(_SCREAMING_SNAKE_CASE ) _construct_solution(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , i - 1 , j - wt[i - 1] , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": lowercase__ = [3, 2, 4, 4] lowercase__ = [4, 3, 2, 3] lowercase__ = 4 lowercase__ = 6 lowercase__ = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] lowercase__ , lowercase__ = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 lowercase__ , lowercase__ = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print('optimal_value = ', optimal_solution) print('An optimal subset corresponding to the optimal value', optimal_subset)
290
"""simple docstring""" class __snake_case : def __init__( self , lowercase , lowercase=None , lowercase=None) -> List[str]: '''simple docstring''' a__: Dict = data a__: List[Any] = previous a__: Any = next_node def __str__( self) -> str: '''simple docstring''' return f'{self.data}' def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.data def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' return self.next def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' return self.previous class __snake_case : def __init__( self , lowercase) -> Dict: '''simple docstring''' a__: List[Any] = head def __iter__( self) -> List[Any]: '''simple docstring''' return self def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if not self.current: raise StopIteration else: a__: Dict = self.current.get_data() a__: Optional[Any] = self.current.get_next() return value class __snake_case : def __init__( self) -> Dict: '''simple docstring''' a__: List[Any] = None # First node in list a__: Optional[int] = None # Last node in list def __str__( self) -> Optional[Any]: '''simple docstring''' a__: Dict = self.head a__: Optional[Any] = [] while current is not None: nodes.append(current.get_data()) a__: str = current.get_next() return " ".join(str(lowercase) for node in nodes) def __contains__( self , lowercase) -> Optional[int]: '''simple docstring''' a__: Optional[int] = self.head while current: if current.get_data() == value: return True a__: Dict = current.get_next() return False def __iter__( self) -> int: '''simple docstring''' return LinkedListIterator(self.head) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' if self.head: return self.head.get_data() return None def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' if self.tail: return self.tail.get_data() return None def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: a__: Optional[Any] = node a__: Optional[Any] = node else: self.insert_before_node(self.head , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' if self.head is None: self.set_head(lowercase) else: self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = Node(lowercase) if self.head is None: self.set_head(lowercase) else: self.set_tail(lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Union[str, Any] = node a__: Optional[Any] = node.previous if node.get_previous() is None: a__: Tuple = node_to_insert else: a__: int = node_to_insert a__: Optional[int] = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Optional[int] = node a__: Tuple = node.next if node.get_next() is None: a__: Optional[int] = node_to_insert else: a__: Any = node_to_insert a__: str = node_to_insert def lowerCamelCase_ ( self , lowercase , lowercase) -> None: '''simple docstring''' a__: Any = 1 a__: Tuple = Node(lowercase) a__: Tuple = self.head while node: if current_position == position: self.insert_before_node(lowercase , lowercase) return current_position += 1 a__: List[Any] = node.next self.insert_after_node(self.tail , lowercase) def lowerCamelCase_ ( self , lowercase) -> Node: '''simple docstring''' a__: Tuple = self.head while node: if node.get_data() == item: return node a__: List[str] = node.get_next() raise Exception('Node not found') def lowerCamelCase_ ( self , lowercase) -> Any: '''simple docstring''' if (node := self.get_node(lowercase)) is not None: if node == self.head: a__: Any = self.head.get_next() if node == self.tail: a__: List[Any] = self.tail.get_previous() self.remove_node_pointers(lowercase) @staticmethod def lowerCamelCase_ ( lowercase) -> None: '''simple docstring''' if node.get_next(): a__: Any = node.previous if node.get_previous(): a__: List[str] = node.next a__: int = None a__: Union[str, Any] = None def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self.head is None def __a ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: Any = '' for word_or_phrase in separated: if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): raise Exception('join() accepts only strings to be joined' ) joined += word_or_phrase + separator return joined.strip(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": from doctest import testmod testmod()
290
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class __snake_case ( __lowerCAmelCase ): a__ = 42 a__ = jnp.floataa a__ = True def lowerCamelCase_ ( self) -> int: '''simple docstring''' super().setup() a__: int = nn.Dense(5 , dtype=self.dtype) def __call__( self , *lowercase , **lowercase) -> Dict: '''simple docstring''' a__: Dict = super().__call__(*lowercase , **lowercase) a__: str = self.cls(outputs[2]) return outputs[:2] + (cls_out,) class __snake_case ( __lowerCAmelCase ): a__ = FlaxBigBirdForNaturalQuestionsModule def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ): a__: Any = logits.shape[-1] a__: List[Any] = (labels[..., None] == jnp.arange(_SCREAMING_SNAKE_CASE )[None]).astype('f4' ) a__: List[str] = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) a__: Dict = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: a__: str = reduction(_SCREAMING_SNAKE_CASE ) return loss a__: Tuple = partial(_SCREAMING_SNAKE_CASE , reduction=jnp.mean ) a__: List[str] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = cross_entropy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class __snake_case : a__ = "google/bigbird-roberta-base" a__ = 3000 a__ = 1_0500 a__ = 128 a__ = 3 a__ = 1 a__ = 5 # tx_args a__ = 3e-5 a__ = 0.0 a__ = 2_0000 a__ = 0.0095 a__ = "bigbird-roberta-natural-questions" a__ = "training-expt" a__ = "data/nq-training.jsonl" a__ = "data/nq-validation.jsonl" def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' os.makedirs(self.base_dir , exist_ok=lowercase) a__: str = os.path.join(self.base_dir , self.save_dir) a__: List[str] = self.batch_size_per_device * jax.device_count() @dataclass class __snake_case : a__ = 42 a__ = 4096 # no dynamic padding on TPUs def __call__( self , lowercase) -> List[Any]: '''simple docstring''' a__: int = self.collate_fn(lowercase) a__: Optional[int] = jax.tree_util.tree_map(lowercase , lowercase) return batch def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__ , a__: Dict = self.fetch_inputs(features['input_ids']) a__: List[Any] = { 'input_ids': jnp.array(lowercase , dtype=jnp.intaa), 'attention_mask': jnp.array(lowercase , dtype=jnp.intaa), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa), } return batch def lowerCamelCase_ ( self , lowercase) -> List[str]: '''simple docstring''' a__: List[Any] = [self._fetch_inputs(lowercase) for ids in input_ids] return zip(*lowercase) def lowerCamelCase_ ( self , lowercase) -> Dict: '''simple docstring''' a__: Union[str, Any] = [1 for _ in range(len(lowercase))] while len(lowercase) < self.max_length: input_ids.append(self.pad_id) attention_mask.append(0) return input_ids, attention_mask def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if seed is not None: a__: int = dataset.shuffle(seed=_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) // batch_size ): a__: Union[str, Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_SCREAMING_SNAKE_CASE ) @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Any: def loss_fn(_SCREAMING_SNAKE_CASE ): a__: str = model_inputs.pop('start_labels' ) a__: Dict = model_inputs.pop('end_labels' ) a__: Optional[int] = model_inputs.pop('pooled_labels' ) a__: Optional[Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=_SCREAMING_SNAKE_CASE , dropout_rng=_SCREAMING_SNAKE_CASE , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: Optional[int] = outputs return state.loss_fn( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) a__ , a__: Union[str, Any] = jax.random.split(_SCREAMING_SNAKE_CASE ) a__: List[Any] = jax.value_and_grad(_SCREAMING_SNAKE_CASE ) a__ , a__: str = grad_fn(state.params ) a__: Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) a__: int = jax.lax.pmean(_SCREAMING_SNAKE_CASE , 'batch' ) a__: Union[str, Any] = state.apply_gradients(grads=_SCREAMING_SNAKE_CASE ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def __a ( _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: Optional[int] = model_inputs.pop('start_labels' ) a__: int = model_inputs.pop('end_labels' ) a__: Dict = model_inputs.pop('pooled_labels' ) a__: Union[str, Any] = state.apply_fn(**_SCREAMING_SNAKE_CASE , params=state.params , train=_SCREAMING_SNAKE_CASE ) a__ , a__ , a__: int = outputs a__: Optional[int] = state.loss_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Tuple = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class __snake_case ( train_state.TrainState ): a__ = struct.field(pytree_node=__lowerCAmelCase ) @dataclass class __snake_case : a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = 42 a__ = None def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase=None) -> Optional[int]: '''simple docstring''' a__: Dict = model.params a__: Any = TrainState.create( apply_fn=model.__call__ , params=lowercase , tx=lowercase , loss_fn=lowercase , ) if ckpt_dir is not None: a__ , a__ , a__ , a__ , a__: Any = restore_checkpoint(lowercase , lowercase) a__: Any = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } a__ , a__: str = build_tx(**lowercase) a__: Optional[Any] = train_state.TrainState( step=lowercase , apply_fn=model.__call__ , params=lowercase , tx=lowercase , opt_state=lowercase , ) a__: int = args a__: Union[str, Any] = data_collator a__: Any = lr a__: Dict = params a__: Tuple = jax_utils.replicate(lowercase) return state def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> int: '''simple docstring''' a__: int = self.args a__: str = len(lowercase) // args.batch_size a__: Tuple = jax.random.PRNGKey(0) a__: List[Any] = jax.random.split(lowercase , jax.device_count()) for epoch in range(args.max_epochs): a__: str = jnp.array(0 , dtype=jnp.floataa) a__: Tuple = get_batched_dataset(lowercase , args.batch_size , seed=lowercase) a__: Optional[int] = 0 for batch in tqdm(lowercase , total=lowercase , desc=f'Running EPOCH-{epoch}'): a__: List[str] = self.data_collator(lowercase) a__ , a__ , a__: int = self.train_step_fn(lowercase , lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 if i % args.logging_steps == 0: a__: List[Any] = jax_utils.unreplicate(state.step) a__: Tuple = running_loss.item() / i a__: Optional[Any] = self.scheduler_fn(state_step - 1) a__: List[Any] = self.evaluate(lowercase , lowercase) a__: List[str] = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(lowercase)) self.logger.log(lowercase , commit=lowercase) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase) -> List[Any]: '''simple docstring''' a__: Tuple = get_batched_dataset(lowercase , self.args.batch_size) a__: Dict = len(lowercase) // self.args.batch_size a__: Tuple = jnp.array(0 , dtype=jnp.floataa) a__: List[Any] = 0 for batch in tqdm(lowercase , total=lowercase , desc='Evaluating ... '): a__: str = self.data_collator(lowercase) a__: List[str] = self.val_step_fn(lowercase , **lowercase) running_loss += jax_utils.unreplicate(metrics['loss']) i += 1 return running_loss / i def lowerCamelCase_ ( self , lowercase , lowercase) -> Any: '''simple docstring''' a__: List[Any] = jax_utils.unreplicate(lowercase) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ') self.model_save_fn(lowercase , params=state.params) with open(os.path.join(lowercase , 'opt_state.msgpack') , 'wb') as f: f.write(to_bytes(state.opt_state)) joblib.dump(self.args , os.path.join(lowercase , 'args.joblib')) joblib.dump(self.data_collator , os.path.join(lowercase , 'data_collator.joblib')) with open(os.path.join(lowercase , 'training_state.json') , 'w') as f: json.dump({'step': state.step.item()} , lowercase) print('DONE') def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'flax_model.msgpack' ) , 'rb' ) as f: a__: int = from_bytes(state.params , f.read() ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'opt_state.msgpack' ) , 'rb' ) as f: a__: Optional[Any] = from_bytes(state.opt_state , f.read() ) a__: Optional[Any] = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'args.joblib' ) ) a__: int = joblib.load(os.path.join(_SCREAMING_SNAKE_CASE , 'data_collator.joblib' ) ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'training_state.json' ) , 'r' ) as f: a__: Any = json.load(_SCREAMING_SNAKE_CASE ) a__: Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: a__: str = num_train_steps - warmup_steps a__: str = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=_SCREAMING_SNAKE_CASE , transition_steps=_SCREAMING_SNAKE_CASE ) a__: List[Any] = optax.linear_schedule(init_value=_SCREAMING_SNAKE_CASE , end_value=1e-7 , transition_steps=_SCREAMING_SNAKE_CASE ) a__: int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Tuple: def weight_decay_mask(_SCREAMING_SNAKE_CASE ): a__: List[Any] = traverse_util.flatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_SCREAMING_SNAKE_CASE ) a__: List[str] = scheduler_fn(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = optax.adamw(learning_rate=_SCREAMING_SNAKE_CASE , weight_decay=_SCREAMING_SNAKE_CASE , mask=_SCREAMING_SNAKE_CASE ) return tx, lr
290
1
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE ) ->str: return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
290
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowercase__ = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __a ( _SCREAMING_SNAKE_CASE ) ->Any: if isinstance(_SCREAMING_SNAKE_CASE , torch.Tensor ): return image elif isinstance(_SCREAMING_SNAKE_CASE , PIL.Image.Image ): a__: Optional[int] = [image] a__: str = [trans(img.convert('RGB' ) ) for img in image] a__: Any = torch.stack(_SCREAMING_SNAKE_CASE ) return image class __snake_case ( __lowerCAmelCase ): def __init__( self , lowercase , lowercase) -> Optional[int]: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM a__: Dict = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowercase , scheduler=lowercase) def lowerCamelCase_ ( self , lowercase) -> int: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f'The value of strength should in [0.0, 1.0] but is {strength}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase) -> Dict: '''simple docstring''' a__: int = min(int(num_inference_steps * strength) , lowercase) a__: Any = max(num_inference_steps - init_timestep , 0) a__: Union[str, Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase=None) -> List[Any]: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( f'`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase)}') a__: Tuple = image.to(device=lowercase , dtype=lowercase) if isinstance(lowercase , lowercase) and len(lowercase) != batch_size: raise ValueError( f'You have passed a list of generators of length {len(lowercase)}, but requested an effective batch' f' size of {batch_size}. Make sure the batch size matches the length of the generators.') a__: List[str] = init_latents.shape a__: List[Any] = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase) # get latents print('add noise to latents at timestep' , lowercase) a__: int = self.scheduler.add_noise(lowercase , lowercase , lowercase) a__: Dict = init_latents return latents @torch.no_grad() def __call__( self , lowercase = None , lowercase = 0.8 , lowercase = 1 , lowercase = None , lowercase = 0.0 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase) # 2. Preprocess image a__: Tuple = preprocess(lowercase) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device) a__ , a__: Union[str, Any] = self.get_timesteps(lowercase , lowercase , self.device) a__: Optional[int] = timesteps[:1].repeat(lowercase) # 4. Prepare latent variables a__: Union[str, Any] = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase) a__: Optional[Any] = latents # 5. Denoising loop for t in self.progress_bar(lowercase): # 1. predict noise model_output a__: Dict = self.unet(lowercase , lowercase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 a__: Optional[Any] = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample a__: Union[str, Any] = (image / 2 + 0.5).clamp(0 , 1) a__: Optional[int] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": a__: Dict = self.numpy_to_pil(lowercase) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase)
290
1
"""simple docstring""" import re from filelock import FileLock try: import nltk lowercase__ = True except (ImportError, ModuleNotFoundError): lowercase__ = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def __a ( _SCREAMING_SNAKE_CASE ) ->str: re.sub('<n>' , '' , _SCREAMING_SNAKE_CASE ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(_SCREAMING_SNAKE_CASE ) )
290
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: Optional[int] = SamImageProcessor() a__: Tuple = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> List[Any]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[Any] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Optional[int] = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: List[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = self.get_image_processor() a__: List[Any] = SamProcessor(image_processor=lowercase) a__: Optional[int] = self.prepare_image_inputs() a__: Optional[Any] = image_processor(lowercase , return_tensors='np') a__: Tuple = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_torch def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: int = self.get_image_processor() a__: List[str] = SamProcessor(image_processor=lowercase) a__: Optional[Any] = [torch.ones((1, 3, 5, 5))] a__: Union[str, Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: int = processor.post_process_masks(lowercase , lowercase , lowercase) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Optional[int] = processor.post_process_masks( lowercase , torch.tensor(lowercase) , torch.tensor(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Dict = [np.ones((1, 3, 5, 5))] a__: Tuple = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = [[1, 0], [0, 1]] with self.assertRaises(lowercase): a__: List[Any] = processor.post_process_masks(lowercase , np.array(lowercase) , np.array(lowercase)) @require_vision @require_tf class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Optional[Any] = tempfile.mkdtemp() a__: List[Any] = SamImageProcessor() a__: Optional[int] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> int: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[Any] = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Optional[int] = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: List[str] = SamProcessor(image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) a__: Dict = self.get_image_processor(do_normalize=lowercase , padding_value=1.0) a__: Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowercase , padding_value=1.0) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , lowercase) def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[Any] = self.get_image_processor() a__: str = SamProcessor(image_processor=lowercase) a__: int = self.prepare_image_inputs() a__: int = image_processor(lowercase , return_tensors='np') a__: Dict = processor(images=lowercase , return_tensors='np') input_feat_extract.pop('original_sizes') # pop original_sizes as it is popped in the processor input_feat_extract.pop('reshaped_input_sizes') # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) @require_tf def lowerCamelCase_ ( self) -> Any: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Any = SamProcessor(image_processor=lowercase) a__: str = [tf.ones((1, 3, 5, 5))] a__: List[Any] = [[17_64, 26_46]] a__: List[Any] = [[6_83, 10_24]] a__: List[Any] = processor.post_process_masks(lowercase , lowercase , lowercase , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: Tuple = processor.post_process_masks( lowercase , tf.convert_to_tensor(lowercase) , tf.convert_to_tensor(lowercase) , return_tensors='tf' , ) self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) # should also work with np a__: Optional[Any] = [np.ones((1, 3, 5, 5))] a__: int = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') self.assertEqual(masks[0].shape , (1, 3, 17_64, 26_46)) a__: List[str] = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError): a__: Any = processor.post_process_masks( lowercase , np.array(lowercase) , np.array(lowercase) , return_tensors='tf') @require_vision @require_torchvision class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' a__: str = tempfile.mkdtemp() a__: int = SamImageProcessor() a__: Union[str, Any] = SamProcessor(lowercase) processor.save_pretrained(self.tmpdirname) def lowerCamelCase_ ( self , **lowercase) -> Optional[int]: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase).image_processor def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Any = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] a__: Any = [Image.fromarray(np.moveaxis(lowercase , 0 , -1)) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Optional[int] = self.get_image_processor() a__: int = SamProcessor(image_processor=lowercase) a__: int = np.random.randint(0 , 2 , size=(1, 3, 5, 5)).astype(np.floataa) a__: Dict = [tf.convert_to_tensor(lowercase)] a__: Union[str, Any] = [torch.tensor(lowercase)] a__: List[Any] = [[17_64, 26_46]] a__: Optional[Any] = [[6_83, 10_24]] a__: Tuple = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='tf') a__: str = processor.post_process_masks( lowercase , lowercase , lowercase , return_tensors='pt') self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy())) @is_pt_tf_cross_test def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Tuple = self.get_image_processor() a__: Dict = SamProcessor(image_processor=lowercase) a__: Any = self.prepare_image_inputs() a__: List[Any] = image_processor(lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Tuple = processor(images=lowercase , return_tensors='pt')['pixel_values'].numpy() a__: Any = image_processor(lowercase , return_tensors='tf')['pixel_values'].numpy() a__: Any = processor(images=lowercase , return_tensors='tf')['pixel_values'].numpy() self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase)) self.assertTrue(np.allclose(lowercase , lowercase))
290
1