code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
_lowerCAmelCase: List[str] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _lowercase( ): a__ =input('Enter message: ' ) a__ =input('Enter key [alphanumeric]: ' ) a__ =input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): a__ ='encrypt' a__ =encrypt_message(__a , __a ) elif mode.lower().startswith('d' ): a__ ='decrypt' a__ =decrypt_message(__a , __a ) print(f"""\n{mode.title()}ed message:""" ) print(__a ) def _lowercase( __a : str , __a : str ): return translate_message(__a , __a , 'encrypt' ) def _lowercase( __a : str , __a : str ): return translate_message(__a , __a , 'decrypt' ) def _lowercase( __a : str , __a : str , __a : str ): a__ =[] a__ =0 a__ =key.upper() for symbol in message: a__ =LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(__a ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(__a ): a__ =0 else: translated.append(__a ) return "".join(__a ) if __name__ == "__main__": main()
20
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: # Initialise PyTorch model _lowercase : Optional[int] = TaConfig.from_json_file(lowerCamelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase : Union[str, Any] = TaForConditionalGeneration(lowerCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
89
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCAmelCase_ : Tuple = logging.get_logger(__name__) class __A ( UpperCamelCase__ ): UpperCamelCase = """upernet""" def __init__( self :int , __snake_case :int=None , __snake_case :Optional[int]=5_12 , __snake_case :Any=0.02 , __snake_case :str=[1, 2, 3, 6] , __snake_case :Optional[Any]=True , __snake_case :Optional[Any]=0.4 , __snake_case :Tuple=3_84 , __snake_case :Optional[int]=2_56 , __snake_case :Dict=1 , __snake_case :Any=False , __snake_case :Tuple=2_55 , **__snake_case :int , ): '''simple docstring''' super().__init__(**__snake_case ) if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) __magic_name__ : Optional[Any] =CONFIG_MAPPING["""resnet"""](out_features=["""stage1""", """stage2""", """stage3""", """stage4"""] ) elif isinstance(__snake_case , __snake_case ): __magic_name__ : int =backbone_config.get("""model_type""" ) __magic_name__ : Optional[Any] =CONFIG_MAPPING[backbone_model_type] __magic_name__ : List[Any] =config_class.from_dict(__snake_case ) __magic_name__ : Dict =backbone_config __magic_name__ : Optional[Any] =hidden_size __magic_name__ : List[str] =initializer_range __magic_name__ : Tuple =pool_scales __magic_name__ : Optional[Any] =use_auxiliary_head __magic_name__ : List[str] =auxiliary_loss_weight __magic_name__ : int =auxiliary_in_channels __magic_name__ : Optional[int] =auxiliary_channels __magic_name__ : Optional[int] =auxiliary_num_convs __magic_name__ : int =auxiliary_concat_input __magic_name__ : Optional[Any] =loss_ignore_index def A__ ( self :Optional[int] ): '''simple docstring''' __magic_name__ : List[str] =copy.deepcopy(self.__dict__ ) __magic_name__ : Dict =self.backbone_config.to_dict() __magic_name__ : Optional[int] =self.__class__.model_type return output
21
def UpperCamelCase_( lowerCamelCase_ ) -> int: if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: _lowercase : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Tuple = 0 _lowercase : List[str] = 2 while digits < n: index += 1 _lowercase : Optional[int] = len(str(fibonacci(lowerCamelCase_ ) ) ) return index def UpperCamelCase_( lowerCamelCase_ = 1000 ) -> int: return fibonacci_digits_index(lowerCamelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
89
0
'''simple docstring''' def snake_case_ (UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float , UpperCamelCase : float , ): '''simple docstring''' _a = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError('''All input parameters must be positive''' ) if any(p > 1 for p in parameters[1:4] ): raise ValueError('''Relative densities cannot be greater than one''' ) else: _a = 1 - (matter_density + radiation_density + dark_energy) _a = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case : List[str] = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
22
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Union[str, Any] = ["ConditionalDetrFeatureExtractor"] SCREAMING_SNAKE_CASE : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Any = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def _snake_case (__lowercase , __lowercase , __lowercase=None , __lowercase=None): if attention_mask is None: UpperCamelCase_ = tf.cast(tf.math.not_equal(__lowercase , config.pad_token_id) , tf.inta) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _a : """simple docstring""" A_ = OPTConfig A_ = {} A_ = """gelu""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase=13 , _UpperCAmelCase=7 , _UpperCAmelCase=True , _UpperCAmelCase=False , _UpperCAmelCase=99 , _UpperCAmelCase=16 , _UpperCAmelCase=2 , _UpperCAmelCase=4 , _UpperCAmelCase=4 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=20 , _UpperCAmelCase=2 , _UpperCAmelCase=1 , _UpperCAmelCase=0 , _UpperCAmelCase=16 , _UpperCAmelCase=16 , ) -> Optional[int]: UpperCamelCase_ = parent UpperCamelCase_ = batch_size UpperCamelCase_ = seq_length UpperCamelCase_ = is_training UpperCamelCase_ = use_labels UpperCamelCase_ = vocab_size UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_act UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = eos_token_id UpperCamelCase_ = pad_token_id UpperCamelCase_ = bos_token_id UpperCamelCase_ = embed_dim UpperCamelCase_ = word_embed_proj_dim UpperCamelCase_ = False def _UpperCAmelCase ( self ) -> Optional[Any]: UpperCamelCase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCamelCase_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCamelCase_ = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCamelCase_ = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=_UpperCAmelCase , **self.config_updates , ) UpperCamelCase_ = prepare_opt_inputs_dict(_UpperCAmelCase , _UpperCAmelCase ) return config, inputs_dict def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: UpperCamelCase_ = TFOPTModel(config=_UpperCAmelCase ) UpperCamelCase_ = inputs_dict['input_ids'] UpperCamelCase_ = input_ids[:1, :] UpperCamelCase_ = inputs_dict['attention_mask'][:1, :] UpperCamelCase_ = 1 # first forward pass UpperCamelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , use_cache=_UpperCAmelCase ) UpperCamelCase_ , UpperCamelCase_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCamelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCamelCase_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCamelCase_ = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCamelCase_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCamelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase )[0] UpperCamelCase_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , past_key_values=_UpperCAmelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCamelCase_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCamelCase_ = output_from_no_past[:, -3:, random_slice_idx] UpperCamelCase_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_UpperCAmelCase , _UpperCAmelCase , rtol=1e-3 ) @require_tf class _a ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): """simple docstring""" A_ = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () A_ = (TFOPTForCausalLM,) if is_tf_available() else () A_ = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) A_ = False A_ = False A_ = False A_ = 10 def _UpperCAmelCase ( self ) -> Optional[Any]: UpperCamelCase_ = TFOPTModelTester(self ) UpperCamelCase_ = ConfigTester(self , config_class=_UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Dict: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> Union[str, Any]: UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Dict: UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(_UpperCAmelCase , _UpperCAmelCase ): if hasattr(_UpperCAmelCase , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(_UpperCAmelCase , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings UpperCamelCase_ = model_class(config=_UpperCAmelCase ) UpperCamelCase_ = _get_word_embedding_weight(_UpperCAmelCase , model.get_input_embeddings() ) UpperCamelCase_ = _get_word_embedding_weight(_UpperCAmelCase , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(_UpperCAmelCase ) UpperCamelCase_ = _get_word_embedding_weight(_UpperCAmelCase , model.get_input_embeddings() ) UpperCamelCase_ = _get_word_embedding_weight(_UpperCAmelCase , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. UpperCamelCase_ = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , _UpperCAmelCase ) # check that weights remain the same after resizing UpperCamelCase_ = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCamelCase_ = False self.assertTrue(_UpperCAmelCase ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , _UpperCAmelCase ) UpperCamelCase_ = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCamelCase_ = False self.assertTrue(_UpperCAmelCase ) def _snake_case (__lowercase): return tf.constant(__lowercase , dtype=tf.intaa) @require_tf class _a ( unittest.TestCase ): """simple docstring""" A_ = 99 def _UpperCAmelCase ( self ) -> List[str]: UpperCamelCase_ = tf.ones((4, 1) , dtype=tf.intaa ) * 2 UpperCamelCase_ = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) UpperCamelCase_ = input_ids.shape[0] UpperCamelCase_ = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _a ( unittest.TestCase ): """simple docstring""" @slow def _UpperCAmelCase ( self ) -> Dict: UpperCamelCase_ = TFOPTModel.from_pretrained('facebook/opt-350m' ) UpperCamelCase_ = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) UpperCamelCase_ = tf.not_equal(_UpperCAmelCase , model.config.pad_token_id ) with tf.GradientTape(): UpperCamelCase_ = model(input_ids=_UpperCAmelCase , attention_mask=_UpperCAmelCase ).last_hidden_state UpperCamelCase_ = (1, 11, 512) self.assertEqual(output.shape , _UpperCAmelCase ) UpperCamelCase_ = tf.constant( [[-0.2_8_7_3, -1.9_2_1_8, -0.3_0_3_3], [-1.2_7_1_0, -0.1_3_3_8, -0.1_9_0_2], [0.4_0_9_5, 0.1_2_1_4, -1.3_1_2_1]] ) self.assertTrue(np.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=4e-3 ) ) UpperCamelCase_ = tf.function(_UpperCAmelCase , jit_compile=_UpperCAmelCase ) UpperCamelCase_ = xla_generate(_UpperCAmelCase , _UpperCAmelCase )[0] self.assertTrue(np.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=4e-2 ) ) @require_tf @slow class _a ( unittest.TestCase ): """simple docstring""" def _UpperCAmelCase ( self ) -> Tuple: super().setUp() UpperCamelCase_ = 'facebook/opt-350m' def _UpperCAmelCase ( self ) -> Optional[int]: UpperCamelCase_ = TFOPTForCausalLM.from_pretrained(self.path_model ) UpperCamelCase_ = GPTaTokenizer.from_pretrained(self.path_model ) UpperCamelCase_ = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False UpperCamelCase_ = tokenizer(_UpperCAmelCase , return_tensors='tf' , padding=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCamelCase_ = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) UpperCamelCase_ = tf.constant( [ [1.3_8_5_1, -1_3.8_9_2_3, -1_0.5_2_2_9, -1_0.7_5_3_3, -0.2_3_0_9, -1_0.2_3_8_4, -0.5_3_6_5, -9.0_9_4_7, -5.1_6_7_0], [-4.7_0_7_3, -1_0.6_2_7_6, -3.9_4_1_5, -2_1.5_2_4_2, -0.2_8_2_2, -0.2_8_2_2, -0.2_8_2_2, -0.2_8_2_2, -0.2_8_2_2], [0.6_2_4_7, -3.4_2_2_9, -8.9_1_7_9, -1.4_2_9_7, -1_4.1_6_5_0, 1.4_1_4_6, -9.0_2_1_8, -0.2_7_0_3, -0.2_7_0_3], [6.4_7_8_3, -1.9_9_1_3, -1_0.7_9_2_6, -2.3_3_3_6, 1.5_0_9_2, -0.9_9_7_4, -6.8_2_1_3, 1.3_4_7_7, 1.3_4_7_7], ] ) self.assertTrue(np.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1e-4 ) ) UpperCamelCase_ = tf.function(_UpperCAmelCase , jit_compile=_UpperCAmelCase ) UpperCamelCase_ = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1e-4 ) ) @require_tf @slow class _a ( unittest.TestCase ): """simple docstring""" @property def _UpperCAmelCase ( self ) -> Any: return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def _UpperCAmelCase ( self ) -> Any: UpperCamelCase_ = 'facebook/opt-125m' UpperCamelCase_ = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] UpperCamelCase_ = [] UpperCamelCase_ = GPTaTokenizer.from_pretrained(_UpperCAmelCase ) UpperCamelCase_ = TFOPTForCausalLM.from_pretrained(_UpperCAmelCase ) for prompt in self.prompts: UpperCamelCase_ = tokenizer(_UpperCAmelCase , return_tensors='tf' ).input_ids UpperCamelCase_ = model.generate(_UpperCAmelCase , max_length=10 ) UpperCamelCase_ = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Union[str, Any]: UpperCamelCase_ = 'facebook/opt-350m' UpperCamelCase_ = GPTaTokenizer.from_pretrained(_UpperCAmelCase ) UpperCamelCase_ = TFOPTForCausalLM.from_pretrained(_UpperCAmelCase ) UpperCamelCase_ = 'left' # use different length sentences to test batching UpperCamelCase_ = [ 'Hello, my dog is a little', 'Today, I', ] UpperCamelCase_ = tokenizer(_UpperCAmelCase , return_tensors='tf' , padding=_UpperCAmelCase ) UpperCamelCase_ = inputs['input_ids'] UpperCamelCase_ = model.generate(input_ids=_UpperCAmelCase , attention_mask=inputs['attention_mask'] ) UpperCamelCase_ = tokenizer(sentences[0] , return_tensors='tf' ).input_ids UpperCamelCase_ = model.generate(input_ids=_UpperCAmelCase ) UpperCamelCase_ = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) UpperCamelCase_ = tokenizer(sentences[1] , return_tensors='tf' ).input_ids UpperCamelCase_ = model.generate(input_ids=_UpperCAmelCase , max_length=model.config.max_length - num_paddings ) UpperCamelCase_ = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) UpperCamelCase_ = tokenizer.decode(output_non_padded[0] , skip_special_tokens=_UpperCAmelCase ) UpperCamelCase_ = tokenizer.decode(output_padded[0] , skip_special_tokens=_UpperCAmelCase ) UpperCamelCase_ = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , [non_padded_sentence, padded_sentence] ) def _UpperCAmelCase ( self ) -> int: UpperCamelCase_ = 'facebook/opt-350m' UpperCamelCase_ = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] UpperCamelCase_ = [] UpperCamelCase_ = GPTaTokenizer.from_pretrained(_UpperCAmelCase ) UpperCamelCase_ = TFOPTForCausalLM.from_pretrained(_UpperCAmelCase ) for prompt in self.prompts: UpperCamelCase_ = tokenizer(_UpperCAmelCase , return_tensors='tf' ).input_ids UpperCamelCase_ = model.generate(_UpperCAmelCase , max_length=10 ) UpperCamelCase_ = tokenizer.batch_decode(_UpperCAmelCase , skip_special_tokens=_UpperCAmelCase ) predicted_outputs += generated_string self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase )
23
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function SCREAMING_SNAKE_CASE : Union[str, Any] = 1.0_5457_1817E-34 # unit of ℏ : J * s SCREAMING_SNAKE_CASE : int = 3E8 # unit of c : m * s^-1 def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> dict[str, float]: if (force, area, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if force < 0: raise ValueError('Magnitude of force can not be negative' ) if distance < 0: raise ValueError('Distance can not be negative' ) if area < 0: raise ValueError('Area can not be negative' ) if force == 0: _lowercase : int = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: _lowercase : List[Any] = (240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: _lowercase : List[Any] = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError('One and only one argument must be 0' ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
89
0
'''simple docstring''' UpperCAmelCase_ : Union[str, Any] = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} UpperCAmelCase_ : str = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def _UpperCamelCase (_lowerCamelCase : dict[int, list[int]] , _lowerCamelCase : int , _lowerCamelCase : list[bool] )-> list[int]: '''simple docstring''' __snake_case = True __snake_case = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) order.append(_lowerCamelCase ) return order def _UpperCamelCase (_lowerCamelCase : dict[int, list[int]] , _lowerCamelCase : int , _lowerCamelCase : list[bool] )-> list[int]: '''simple docstring''' __snake_case = True __snake_case = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) return component def _UpperCamelCase (_lowerCamelCase : dict[int, list[int]] )-> list[list[int]]: '''simple docstring''' __snake_case = len(_lowerCamelCase ) * [False] __snake_case = {vert: [] for vert in range(len(_lowerCamelCase ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(_lowerCamelCase ) __snake_case = [] for i, was_visited in enumerate(_lowerCamelCase ): if not was_visited: order += topology_sort(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) __snake_case = [] __snake_case = len(_lowerCamelCase ) * [False] for i in range(len(_lowerCamelCase ) ): __snake_case = order[len(_lowerCamelCase ) - i - 1] if not visited[vert]: __snake_case = find_components(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) components_list.append(_lowerCamelCase ) return components_list
24
def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) _lowercase : List[str] = 0 _lowercase : Optional[int] = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Any = [int(lowerCamelCase_ ) for i in num_string] _lowercase : List[Any] = 1 for i in range(0 , len(lowerCamelCase_ ) ): total *= numbers[i] _lowercase : Optional[Any] = str(lowerCamelCase_ ) steps += 1 return steps def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) _lowercase : Optional[int] = 0 _lowercase : str = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Dict = [int(lowerCamelCase_ ) for i in num_string] _lowercase : Any = 0 for i in range(0 , len(lowerCamelCase_ ) ): total += numbers[i] _lowercase : Dict = str(lowerCamelCase_ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
89
0
import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def lowerCamelCase__ ( _a): random.seed(_a) np.random.seed(_a) torch.manual_seed(_a) torch.cuda.manual_seed_all(_a) # ^^ safe to call this function even if cuda is not available class _UpperCamelCase : '''simple docstring''' def __init__( self : List[str] , a : Iterable[torch.nn.Parameter] , a : float = 0.9999 , a : float = 0.0 , a : int = 0 , a : bool = False , a : Union[float, int] = 1.0 , a : Union[float, int] = 2 / 3 , a : Optional[Any] = None , a : Dict[str, Any] = None , **a : Dict , ) -> Tuple: """simple docstring""" if isinstance(a , torch.nn.Module ): SCREAMING_SNAKE_CASE : Optional[Any] = ( "Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. " "Please pass the parameters of the module instead." ) deprecate( "passing a `torch.nn.Module` to `ExponentialMovingAverage`" , "1.0.0" , a , standard_warn=a , ) SCREAMING_SNAKE_CASE : Optional[Any] = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility SCREAMING_SNAKE_CASE : Tuple = True if kwargs.get("max_value" , a ) is not None: SCREAMING_SNAKE_CASE : List[Any] = "The `max_value` argument is deprecated. Please use `decay` instead." deprecate("max_value" , "1.0.0" , a , standard_warn=a ) SCREAMING_SNAKE_CASE : str = kwargs["max_value"] if kwargs.get("min_value" , a ) is not None: SCREAMING_SNAKE_CASE : Union[str, Any] = "The `min_value` argument is deprecated. Please use `min_decay` instead." deprecate("min_value" , "1.0.0" , a , standard_warn=a ) SCREAMING_SNAKE_CASE : List[Any] = kwargs["min_value"] SCREAMING_SNAKE_CASE : Optional[int] = list(a ) SCREAMING_SNAKE_CASE : str = [p.clone().detach() for p in parameters] if kwargs.get("device" , a ) is not None: SCREAMING_SNAKE_CASE : int = "The `device` argument is deprecated. Please use `to` instead." deprecate("device" , "1.0.0" , a , standard_warn=a ) self.to(device=kwargs["device"] ) SCREAMING_SNAKE_CASE : Union[str, Any] = None SCREAMING_SNAKE_CASE : Optional[int] = decay SCREAMING_SNAKE_CASE : Tuple = min_decay SCREAMING_SNAKE_CASE : Tuple = update_after_step SCREAMING_SNAKE_CASE : Tuple = use_ema_warmup SCREAMING_SNAKE_CASE : List[Any] = inv_gamma SCREAMING_SNAKE_CASE : int = power SCREAMING_SNAKE_CASE : Optional[int] = 0 SCREAMING_SNAKE_CASE : List[Any] = None # set in `step()` SCREAMING_SNAKE_CASE : Union[str, Any] = model_cls SCREAMING_SNAKE_CASE : Dict = model_config @classmethod def __UpperCamelCase ( cls : Optional[Any] , a : Optional[int] , a : Tuple ) -> "EMAModel": """simple docstring""" SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Tuple = model_cls.load_config(a , return_unused_kwargs=a ) SCREAMING_SNAKE_CASE : Any = model_cls.from_pretrained(a ) SCREAMING_SNAKE_CASE : Tuple = cls(model.parameters() , model_cls=a , model_config=model.config ) ema_model.load_state_dict(a ) return ema_model def __UpperCamelCase ( self : List[str] , a : Any ) -> List[Any]: """simple docstring""" if self.model_cls is None: raise ValueError("`save_pretrained` can only be used if `model_cls` was defined at __init__." ) if self.model_config is None: raise ValueError("`save_pretrained` can only be used if `model_config` was defined at __init__." ) SCREAMING_SNAKE_CASE : List[Any] = self.model_cls.from_config(self.model_config ) SCREAMING_SNAKE_CASE : List[str] = self.state_dict() state_dict.pop("shadow_params" , a ) model.register_to_config(**a ) self.copy_to(model.parameters() ) model.save_pretrained(a ) def __UpperCamelCase ( self : List[str] , a : int ) -> float: """simple docstring""" SCREAMING_SNAKE_CASE : int = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: SCREAMING_SNAKE_CASE : Any = 1 - (1 + step / self.inv_gamma) ** -self.power else: SCREAMING_SNAKE_CASE : List[str] = (1 + step) / (10 + step) SCREAMING_SNAKE_CASE : Optional[Any] = min(a , self.decay ) # make sure decay is not smaller than min_decay SCREAMING_SNAKE_CASE : str = max(a , self.min_decay ) return cur_decay_value @torch.no_grad() def __UpperCamelCase ( self : Optional[int] , a : Iterable[torch.nn.Parameter] ) -> int: """simple docstring""" if isinstance(a , torch.nn.Module ): SCREAMING_SNAKE_CASE : Tuple = ( "Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. " "Please pass the parameters of the module instead." ) deprecate( "passing a `torch.nn.Module` to `ExponentialMovingAverage.step`" , "1.0.0" , a , standard_warn=a , ) SCREAMING_SNAKE_CASE : List[Any] = parameters.parameters() SCREAMING_SNAKE_CASE : List[str] = list(a ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. SCREAMING_SNAKE_CASE : Union[str, Any] = self.get_decay(self.optimization_step ) SCREAMING_SNAKE_CASE : Any = decay SCREAMING_SNAKE_CASE : Union[str, Any] = 1 - decay SCREAMING_SNAKE_CASE : Union[str, Any] = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , a ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): SCREAMING_SNAKE_CASE : List[str] = deepspeed.zero.GatheredParameters(a , modifier_rank=a ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(a ) def __UpperCamelCase ( self : Optional[Any] , a : Iterable[torch.nn.Parameter] ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : Any = list(a ) for s_param, param in zip(self.shadow_params , a ): param.data.copy_(s_param.to(param.device ).data ) def __UpperCamelCase ( self : List[str] , a : int=None , a : Union[str, Any]=None ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : int = [ p.to(device=a , dtype=a ) if p.is_floating_point() else p.to(device=a ) for p in self.shadow_params ] def __UpperCamelCase ( self : Dict ) -> dict: """simple docstring""" return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def __UpperCamelCase ( self : Tuple , a : Iterable[torch.nn.Parameter] ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = [param.detach().cpu().clone() for param in parameters] def __UpperCamelCase ( self : Tuple , a : Iterable[torch.nn.Parameter] ) -> None: """simple docstring""" if self.temp_stored_params is None: raise RuntimeError("This ExponentialMovingAverage has no `store()`ed weights " "to `restore()`" ) for c_param, param in zip(self.temp_stored_params , a ): param.data.copy_(c_param.data ) # Better memory-wise. SCREAMING_SNAKE_CASE : Tuple = None def __UpperCamelCase ( self : Optional[int] , a : dict ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : int = copy.deepcopy(a ) SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.get("decay" , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError("Decay must be between 0 and 1" ) SCREAMING_SNAKE_CASE : Optional[int] = state_dict.get("min_decay" , self.min_decay ) if not isinstance(self.min_decay , a ): raise ValueError("Invalid min_decay" ) SCREAMING_SNAKE_CASE : List[str] = state_dict.get("optimization_step" , self.optimization_step ) if not isinstance(self.optimization_step , a ): raise ValueError("Invalid optimization_step" ) SCREAMING_SNAKE_CASE : Dict = state_dict.get("update_after_step" , self.update_after_step ) if not isinstance(self.update_after_step , a ): raise ValueError("Invalid update_after_step" ) SCREAMING_SNAKE_CASE : List[Any] = state_dict.get("use_ema_warmup" , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , a ): raise ValueError("Invalid use_ema_warmup" ) SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.get("inv_gamma" , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError("Invalid inv_gamma" ) SCREAMING_SNAKE_CASE : List[Any] = state_dict.get("power" , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError("Invalid power" ) SCREAMING_SNAKE_CASE : Tuple = state_dict.get("shadow_params" , a ) if shadow_params is not None: SCREAMING_SNAKE_CASE : List[Any] = shadow_params if not isinstance(self.shadow_params , a ): raise ValueError("shadow_params must be a list" ) if not all(isinstance(a , torch.Tensor ) for p in self.shadow_params ): raise ValueError("shadow_params must all be Tensors" )
25
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: # initialize config if "resnet-50" in model_name: _lowercase : Union[str, Any] = ResNetConfig.from_pretrained('microsoft/resnet-50' ) elif "resnet-101" in model_name: _lowercase : Optional[Any] = ResNetConfig.from_pretrained('microsoft/resnet-101' ) else: raise ValueError('Model name should include either resnet50 or resnet101' ) _lowercase : Tuple = DetrConfig(use_timm_backbone=lowerCamelCase_ , backbone_config=lowerCamelCase_ ) # set label attributes _lowercase : Any = 'panoptic' in model_name if is_panoptic: _lowercase : List[Any] = 250 else: _lowercase : str = 91 _lowercase : List[Any] = 'huggingface/label-files' _lowercase : Any = 'coco-detection-id2label.json' _lowercase : Tuple = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : int = idalabel _lowercase : Any = {v: k for k, v in idalabel.items()} return config, is_panoptic def UpperCamelCase_( lowerCamelCase_ ) -> Any: # here we list all keys to be renamed (original name on the left, our name on the right) _lowercase : List[str] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.conv1.weight', 'backbone.conv_encoder.model.embedder.embedder.convolution.weight') ) rename_keys.append(('backbone.0.body.bn1.weight', 'backbone.conv_encoder.model.embedder.embedder.normalization.weight') ) rename_keys.append(('backbone.0.body.bn1.bias', 'backbone.conv_encoder.model.embedder.embedder.normalization.bias') ) rename_keys.append(('backbone.0.body.bn1.running_mean', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_mean') ) rename_keys.append(('backbone.0.body.bn1.running_var', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_var') ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var''', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var''', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ] ) return rename_keys def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : str = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> str: _lowercase : Any = '' if is_panoptic: _lowercase : Optional[Any] = 'detr.' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowercase : int = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Tuple = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[str] = in_proj_weight[:256, :] _lowercase : Tuple = in_proj_bias[:256] _lowercase : List[Any] = in_proj_weight[256:512, :] _lowercase : Any = in_proj_bias[256:512] _lowercase : int = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _lowercase : str = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Optional[int] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : Union[str, Any] = in_proj_weight[:256, :] _lowercase : Dict = in_proj_bias[:256] _lowercase : Tuple = in_proj_weight[256:512, :] _lowercase : Dict = in_proj_bias[256:512] _lowercase : str = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _lowercase : Tuple = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) _lowercase : Dict = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _lowercase : List[str] = in_proj_weight_cross_attn[:256, :] _lowercase : Tuple = in_proj_bias_cross_attn[:256] _lowercase : str = in_proj_weight_cross_attn[256:512, :] _lowercase : Union[str, Any] = in_proj_bias_cross_attn[256:512] _lowercase : List[Any] = in_proj_weight_cross_attn[-256:, :] _lowercase : Dict = in_proj_bias_cross_attn[-256:] def UpperCamelCase_( ) -> List[Any]: _lowercase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : str = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) return im @torch.no_grad() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=False ) -> List[Any]: _lowercase , _lowercase : int = get_detr_config(lowerCamelCase_ ) # load original model from torch hub _lowercase : int = { 'detr-resnet-50': 'detr_resnet50', 'detr-resnet-101': 'detr_resnet101', } logger.info(F'''Converting model {model_name}...''' ) _lowercase : Optional[Any] = torch.hub.load('facebookresearch/detr' , model_name_to_original_name[model_name] , pretrained=lowerCamelCase_ ).eval() _lowercase : str = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCamelCase_ ): if is_panoptic: _lowercase : str = 'detr.' + src rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCamelCase_ , is_panoptic=lowerCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowercase : List[Any] = 'detr.model.' if is_panoptic else 'model.' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('detr' ) and not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ) ): _lowercase : Tuple = state_dict.pop(lowerCamelCase_ ) _lowercase : int = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val elif key.startswith('bbox_attention' ) or key.startswith('mask_head' ): continue else: _lowercase : Optional[Any] = state_dict.pop(lowerCamelCase_ ) _lowercase : Union[str, Any] = val else: if not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : List[str] = val # finally, create HuggingFace model and load state dict _lowercase : Optional[Any] = DetrForSegmentation(lowerCamelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCamelCase_ ) model.load_state_dict(lowerCamelCase_ ) model.eval() # verify our conversion on an image _lowercase : str = 'coco_panoptic' if is_panoptic else 'coco_detection' _lowercase : Optional[int] = DetrImageProcessor(format=lowerCamelCase_ ) _lowercase : str = processor(images=prepare_img() , return_tensors='pt' ) _lowercase : Tuple = encoding['pixel_values'] _lowercase : int = detr(lowerCamelCase_ ) _lowercase : Tuple = model(lowerCamelCase_ ) assert torch.allclose(outputs.logits , original_outputs['pred_logits'] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs['pred_boxes'] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['pred_masks'] , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info('Uploading PyTorch model and image processor to the hub...' ) model.push_to_hub(F'''nielsr/{model_name}''' ) processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_name", default="detr-resnet-50", type=str, choices=["detr-resnet-50", "detr-resnet-101"], help="Name of the DETR model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the model to the hub or not.") SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
'''simple docstring''' import os import sys __UpperCamelCase = os.path.join(os.path.dirname(__file__), "src") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) __UpperCamelCase = [ "torch", "numpy", "tokenizers", "filelock", "requests", "tqdm", "regex", "sentencepiece", "sacremoses", "importlib_metadata", "huggingface_hub", ] @add_start_docstrings(AutoConfig.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> Tuple: """simple docstring""" return AutoConfig.from_pretrained(*_lowerCamelCase , **_lowerCamelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> List[Any]: """simple docstring""" return AutoTokenizer.from_pretrained(*_lowerCamelCase , **_lowerCamelCase ) @add_start_docstrings(AutoModel.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> Union[str, Any]: """simple docstring""" return AutoModel.from_pretrained(*_lowerCamelCase , **_lowerCamelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> Dict: """simple docstring""" return AutoModelForCausalLM.from_pretrained(*_lowerCamelCase , **_lowerCamelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> Dict: """simple docstring""" return AutoModelForMaskedLM.from_pretrained(*_lowerCamelCase , **_lowerCamelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> Union[str, Any]: """simple docstring""" return AutoModelForSequenceClassification.from_pretrained(*_lowerCamelCase , **_lowerCamelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def _a ( *_lowerCamelCase , **_lowerCamelCase ) -> Optional[int]: """simple docstring""" return AutoModelForQuestionAnswering.from_pretrained(*_lowerCamelCase , **_lowerCamelCase )
26
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput SCREAMING_SNAKE_CASE : str = "scheduler_config.json" class _lowerCamelCase( _a ): lowercase_ : Any = 1 lowercase_ : Dict = 2 lowercase_ : Union[str, Any] = 3 lowercase_ : Tuple = 4 lowercase_ : Optional[Any] = 5 @dataclass class _lowerCamelCase( _a ): lowercase_ : jnp.ndarray class _lowerCamelCase: lowercase_ : Union[str, Any] = SCHEDULER_CONFIG_NAME lowercase_ : str = ["""dtype"""] lowercase_ : Dict = [] lowercase_ : int = True @classmethod def UpperCamelCase ( cls, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase=False, **lowerCamelCase, ) -> Optional[int]: """simple docstring""" _lowercase , _lowercase : Optional[int] = cls.load_config( pretrained_model_name_or_path=lowerCamelCase, subfolder=lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase, ) _lowercase , _lowercase : Tuple = cls.from_config(lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase) if hasattr(lowerCamelCase, 'create_state') and getattr(lowerCamelCase, 'has_state', lowerCamelCase): _lowercase : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, **lowerCamelCase) -> Any: """simple docstring""" self.save_config(save_directory=lowerCamelCase, push_to_hub=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return self._get_compatibles() @classmethod def UpperCamelCase ( cls) -> Any: """simple docstring""" _lowercase : Any = list(set([cls.__name__] + cls._compatibles)) _lowercase : Dict = importlib.import_module(__name__.split('.')[0]) _lowercase : Any = [ getattr(lowerCamelCase, lowerCamelCase) for c in compatible_classes_str if hasattr(lowerCamelCase, lowerCamelCase) ] return compatible_classes def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> jnp.ndarray: assert len(lowerCamelCase_ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(lowerCamelCase_ ) - x.ndim) ) , lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=0.9_99 , lowerCamelCase_=jnp.floataa ) -> jnp.ndarray: def alpha_bar(lowerCamelCase_ ): return math.cos((time_step + 0.0_08) / 1.0_08 * math.pi / 2 ) ** 2 _lowercase : List[Any] = [] for i in range(lowerCamelCase_ ): _lowercase : Any = i / num_diffusion_timesteps _lowercase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(lowerCamelCase_ ) / alpha_bar(lowerCamelCase_ ) , lowerCamelCase_ ) ) return jnp.array(lowerCamelCase_ , dtype=lowerCamelCase_ ) @flax.struct.dataclass class _lowerCamelCase: lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray @classmethod def UpperCamelCase ( cls, lowerCamelCase) -> str: """simple docstring""" _lowercase : int = scheduler.config if config.trained_betas is not None: _lowercase : str = jnp.asarray(config.trained_betas, dtype=scheduler.dtype) elif config.beta_schedule == "linear": _lowercase : List[Any] = jnp.linspace(config.beta_start, config.beta_end, config.num_train_timesteps, dtype=scheduler.dtype) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _lowercase : Dict = ( jnp.linspace( config.beta_start**0.5, config.beta_end**0.5, config.num_train_timesteps, dtype=scheduler.dtype) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _lowercase : Optional[int] = betas_for_alpha_bar(config.num_train_timesteps, dtype=scheduler.dtype) else: raise NotImplementedError( F'''beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}''') _lowercase : List[str] = 1.0 - betas _lowercase : Union[str, Any] = jnp.cumprod(lowerCamelCase, axis=0) return cls( alphas=lowerCamelCase, betas=lowerCamelCase, alphas_cumprod=lowerCamelCase, ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : str = state.alphas_cumprod _lowercase : str = alphas_cumprod[timesteps] ** 0.5 _lowercase : Optional[Any] = sqrt_alpha_prod.flatten() _lowercase : Tuple = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) _lowercase : Union[str, Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowercase : Optional[Any] = sqrt_one_minus_alpha_prod.flatten() _lowercase : int = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase , _lowercase : Optional[int] = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: _lowercase , _lowercase : Tuple = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
89
0
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowerCamelCase( __snake_case , unittest.TestCase ): '''simple docstring''' __magic_name__ = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def lowerCAmelCase__ ( self , snake_case_=0 ): _A = np.random.RandomState(snake_case_ ) _A = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = pipe(**snake_case_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _A = np.array([0.6_5072, 0.5_8492, 0.4_8219, 0.5_5521, 0.5_3180, 0.5_5939, 0.5_0697, 0.3_9800, 0.4_6455] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) _A = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = pipe(**snake_case_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _A = np.array([0.6_5863, 0.5_9425, 0.4_9326, 0.5_6313, 0.5_3875, 0.5_6627, 0.5_1065, 0.3_9777, 0.4_6330] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) _A = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = pipe(**snake_case_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _A = np.array([0.5_3755, 0.6_0786, 0.4_7402, 0.4_9488, 0.5_1869, 0.4_9819, 0.4_7985, 0.3_8957, 0.4_4279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) _A = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = pipe(**snake_case_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _A = np.array([0.5_3755, 0.6_0786, 0.4_7402, 0.4_9488, 0.5_1869, 0.4_9819, 0.4_7985, 0.3_8957, 0.4_4279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) _A = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = pipe(**snake_case_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _A = np.array([0.5_3817, 0.6_0812, 0.4_7384, 0.4_9530, 0.5_1894, 0.4_9814, 0.4_7984, 0.3_8958, 0.4_4271] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) _A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = pipe(**snake_case_ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _A = np.array([0.5_3895, 0.6_0808, 0.4_7933, 0.4_9608, 0.5_1886, 0.4_9950, 0.4_8053, 0.3_8957, 0.4_4200] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = 3 * [inputs['prompt']] # forward _A = pipe(**snake_case_ ) _A = output.images[0, -3:, -3:, -1] _A = self.get_dummy_inputs() _A = 3 * [inputs.pop('prompt' )] _A = pipe.tokenizer( snake_case_ , padding='max_length' , max_length=pipe.tokenizer.model_max_length , truncation=snake_case_ , return_tensors='np' , ) _A = text_inputs['input_ids'] _A = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] _A = prompt_embeds # forward _A = pipe(**snake_case_ ) _A = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = self.get_dummy_inputs() _A = 3 * ['this is a negative prompt'] _A = negative_prompt _A = 3 * [inputs['prompt']] # forward _A = pipe(**snake_case_ ) _A = output.images[0, -3:, -3:, -1] _A = self.get_dummy_inputs() _A = 3 * [inputs.pop('prompt' )] _A = [] for p in [prompt, negative_prompt]: _A = pipe.tokenizer( snake_case_ , padding='max_length' , max_length=pipe.tokenizer.model_max_length , truncation=snake_case_ , return_tensors='np' , ) _A = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) _A, _A = embeds # forward _A = pipe(**snake_case_ ) _A = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class lowerCamelCase( unittest.TestCase ): '''simple docstring''' @property def lowerCAmelCase__ ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCAmelCase__ ( self ): _A = ort.SessionOptions() _A = False return options def lowerCAmelCase__ ( self ): # using the PNDM scheduler by default _A = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='onnx' , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'A painting of a squirrel eating a burger' np.random.seed(0 ) _A = sd_pipe([prompt] , guidance_scale=6.0 , num_inference_steps=10 , output_type='np' ) _A = output.images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _A = np.array([0.0452, 0.0390, 0.0087, 0.0350, 0.0617, 0.0364, 0.0544, 0.0523, 0.0720] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCAmelCase__ ( self ): _A = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5' , subfolder='scheduler' , revision='onnx' ) _A = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , revision='onnx' , scheduler=snake_case_ , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'open neural network exchange' _A = np.random.RandomState(0 ) _A = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case_ , output_type='np' ) _A = output.images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _A = np.array([0.2867, 0.1974, 0.1481, 0.7294, 0.7251, 0.6667, 0.4194, 0.5642, 0.6486] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCAmelCase__ ( self ): _A = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5' , subfolder='scheduler' , revision='onnx' ) _A = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , revision='onnx' , scheduler=snake_case_ , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'open neural network exchange' _A = np.random.RandomState(0 ) _A = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case_ , output_type='np' ) _A = output.images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _A = np.array([0.2306, 0.1959, 0.1593, 0.6549, 0.6394, 0.5408, 0.5065, 0.6010, 0.6161] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowerCAmelCase__ ( self ): _A = 0 def test_callback_fn(snake_case_ , snake_case_ , snake_case_ ) -> None: _A = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _A = latents[0, -3:, -3:, -1] _A = np.array( [-0.6772, -0.3835, -1.2456, 0.1905, -1.0974, 0.6967, -1.9353, 0.0178, 1.0167] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _A = latents[0, -3:, -3:, -1] _A = np.array( [-0.3351, 0.2241, -0.1837, -0.2325, -0.6577, 0.3393, -0.0241, 0.5899, 1.3875] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 _A = False _A = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , revision='onnx' , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case_ ) _A = 'Andromeda galaxy in a bottle' _A = np.random.RandomState(0 ) pipe( prompt=snake_case_ , num_inference_steps=5 , guidance_scale=7.5 , generator=snake_case_ , callback=snake_case_ , callback_steps=1 , ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def lowerCAmelCase__ ( self ): _A = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , revision='onnx' , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) assert isinstance(snake_case_ , snake_case_ ) assert pipe.safety_checker is None _A = pipe('example prompt' , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(snake_case_ ) _A = OnnxStableDiffusionPipeline.from_pretrained(snake_case_ ) # sanity check that the pipeline still works assert pipe.safety_checker is None _A = pipe('example prompt' , num_inference_steps=2 ).images[0] assert image is not None
27
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> float: if not nums: raise ValueError('List is empty' ) return sum(lowerCamelCase_ ) / len(lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
89
0
'''simple docstring''' from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging UpperCamelCase_ = logging.get_logger(__name__) class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' A : List[str] = ['''pixel_values'''] def __init__( self, A = True, A = None, A = PILImageResampling.BICUBIC, A = True, A = None, A = True, A = 1 / 255, A = True, A = IMAGENET_DEFAULT_MEAN, A = IMAGENET_DEFAULT_STD, **A, ): '''simple docstring''' super().__init__(**A ) SCREAMING_SNAKE_CASE : Any = size if size is not None else {'shortest_edge': 224} SCREAMING_SNAKE_CASE : List[str] = get_size_dict(A, default_to_square=A ) SCREAMING_SNAKE_CASE : Tuple = crop_size if crop_size is not None else {'height': 224, 'width': 224} SCREAMING_SNAKE_CASE : List[str] = get_size_dict(A, param_name='crop_size' ) SCREAMING_SNAKE_CASE : Any = do_resize SCREAMING_SNAKE_CASE : List[str] = size SCREAMING_SNAKE_CASE : Any = resample SCREAMING_SNAKE_CASE : Optional[Any] = do_center_crop SCREAMING_SNAKE_CASE : Any = crop_size SCREAMING_SNAKE_CASE : Tuple = do_rescale SCREAMING_SNAKE_CASE : Optional[Any] = rescale_factor SCREAMING_SNAKE_CASE : List[Any] = do_normalize SCREAMING_SNAKE_CASE : int = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN SCREAMING_SNAKE_CASE : List[str] = image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCamelCase_ ( self, A, A, A = PILImageResampling.BICUBIC, A = None, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = get_size_dict(A, default_to_square=A ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: SCREAMING_SNAKE_CASE : str = int((256 / 224) * size['shortest_edge'] ) SCREAMING_SNAKE_CASE : str = get_resize_output_image_size(A, size=A, default_to_square=A ) SCREAMING_SNAKE_CASE : Dict = {'height': output_size[0], 'width': output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}" ) return resize( A, size=(size_dict['height'], size_dict['width']), resample=A, data_format=A, **A ) def UpperCamelCase_ ( self, A, A, A = None, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(F"Size dict must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(A, size=(size['height'], size['width']), data_format=A, **A ) def UpperCamelCase_ ( self, A, A, A = None, **A, ): '''simple docstring''' return rescale(A, scale=A, data_format=A, **A ) def UpperCamelCase_ ( self, A, A, A, A = None, **A, ): '''simple docstring''' return normalize(A, mean=A, std=A, data_format=A, **A ) def UpperCamelCase_ ( self, A, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = None, A = ChannelDimension.FIRST, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE : Optional[Any] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE : List[str] = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE : Optional[Any] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE : Optional[Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE : Dict = size if size is not None else self.size SCREAMING_SNAKE_CASE : List[str] = get_size_dict(A, default_to_square=A ) SCREAMING_SNAKE_CASE : Tuple = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE : Union[str, Any] = get_size_dict(A, param_name='crop_size' ) SCREAMING_SNAKE_CASE : int = make_list_of_images(A ) if not valid_images(A ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE : Tuple = [to_numpy_array(A ) for image in images] if do_resize: SCREAMING_SNAKE_CASE : List[str] = [self.resize(A, A, A ) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE : int = [self.center_crop(A, A ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE : int = [self.rescale(A, A ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE : List[str] = [self.normalize(A, A, A ) for image in images] SCREAMING_SNAKE_CASE : Optional[int] = [to_channel_dimension_format(A, A ) for image in images] SCREAMING_SNAKE_CASE : Tuple = {'pixel_values': images} return BatchFeature(data=A, tensor_type=A )
28
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def UpperCamelCase_( ) -> List[Any]: _lowercase : int = ArgumentParser('Transformers CLI tool' , usage='transformers-cli <command> [<args>]' ) _lowercase : Optional[Any] = parser.add_subparsers(help='transformers-cli command helpers' ) # Register commands ConvertCommand.register_subcommand(lowerCamelCase_ ) DownloadCommand.register_subcommand(lowerCamelCase_ ) EnvironmentCommand.register_subcommand(lowerCamelCase_ ) RunCommand.register_subcommand(lowerCamelCase_ ) ServeCommand.register_subcommand(lowerCamelCase_ ) UserCommands.register_subcommand(lowerCamelCase_ ) AddNewModelCommand.register_subcommand(lowerCamelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCamelCase_ ) LfsCommands.register_subcommand(lowerCamelCase_ ) PTtoTFCommand.register_subcommand(lowerCamelCase_ ) # Let's go _lowercase : Any = parser.parse_args() if not hasattr(lowerCamelCase_ , 'func' ): parser.print_help() exit(1 ) # Run _lowercase : Optional[int] = args.func(lowerCamelCase_ ) service.run() if __name__ == "__main__": main()
89
0
"""simple docstring""" from jiwer import compute_measures import datasets A_ = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ A_ = """\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. """ A_ = """ Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> wer = datasets.load_metric(\"wer\") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCamelCase ( datasets.Metric ): def UpperCAmelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ] , ) def UpperCAmelCase__ ( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=False ): if concatenate_texts: return compute_measures(UpperCAmelCase , UpperCAmelCase )["wer"] else: lowerCamelCase_ = 0 lowerCamelCase_ = 0 for prediction, reference in zip(UpperCAmelCase , UpperCAmelCase ): lowerCamelCase_ = compute_measures(UpperCAmelCase , UpperCAmelCase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
29
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Optional[int] = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def UpperCamelCase ( self, lowerCamelCase=0) -> str: """simple docstring""" _lowercase : Optional[int] = np.random.RandomState(lowerCamelCase) _lowercase : Union[str, Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : List[Any] = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[str] = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Any = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_dummy_inputs() _lowercase : Any = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : Optional[int] = output.images[0, -3:, -3:, -1] _lowercase : int = self.get_dummy_inputs() _lowercase : Union[str, Any] = 3 * [inputs.pop('prompt')] _lowercase : Union[str, Any] = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Tuple = text_inputs['input_ids'] _lowercase : Any = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0] _lowercase : List[Any] = prompt_embeds # forward _lowercase : Union[str, Any] = pipe(**lowerCamelCase) _lowercase : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs() _lowercase : Any = 3 * ['this is a negative prompt'] _lowercase : str = negative_prompt _lowercase : Optional[int] = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : str = output.images[0, -3:, -3:, -1] _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : str = 3 * [inputs.pop('prompt')] _lowercase : Optional[int] = [] for p in [prompt, negative_prompt]: _lowercase : Tuple = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Dict = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0]) _lowercase , _lowercase : str = embeds # forward _lowercase : Dict = pipe(**lowerCamelCase) _lowercase : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = ort.SessionOptions() _lowercase : str = False return options def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'A painting of a squirrel eating a burger' np.random.seed(0) _lowercase : Union[str, Any] = sd_pipe([prompt], guidance_scale=6.0, num_inference_steps=10, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : str = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'open neural network exchange' _lowercase : List[Any] = np.random.RandomState(0) _lowercase : Optional[Any] = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = 'open neural network exchange' _lowercase : str = np.random.RandomState(0) _lowercase : Dict = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[Any] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : List[Any] = 0 def test_callback_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase) -> None: _lowercase : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _lowercase : Any = latents[0, -3:, -3:, -1] _lowercase : Tuple = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _lowercase : List[Any] = latents[0, -3:, -3:, -1] _lowercase : str = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 _lowercase : Any = False _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = 'Andromeda galaxy in a bottle' _lowercase : str = np.random.RandomState(0) pipe( prompt=lowerCamelCase, num_inference_steps=5, guidance_scale=7.5, generator=lowerCamelCase, callback=lowerCamelCase, callback_steps=1, ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) assert isinstance(lowerCamelCase, lowerCamelCase) assert pipe.safety_checker is None _lowercase : Optional[int] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase) _lowercase : Any = OnnxStableDiffusionPipeline.from_pretrained(lowerCamelCase) # sanity check that the pipeline still works assert pipe.safety_checker is None _lowercase : List[str] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None
89
0
def lowerCamelCase__ ( _lowercase , _lowercase ): '''simple docstring''' if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
30
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : List[Any] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = ["PoolFormerFeatureExtractor"] SCREAMING_SNAKE_CASE : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[Any] = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
89
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = ["image_processor", "tokenizer"] lowercase_ = "CLIPImageProcessor" lowercase_ = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self : Tuple , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[Any]=None , **_lowerCAmelCase : Any ): SCREAMING_SNAKE_CASE_ = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , _lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ = kwargs.pop('feature_extractor' ) SCREAMING_SNAKE_CASE_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(_lowerCAmelCase , _lowerCAmelCase ) def __call__( self : Tuple , _lowerCAmelCase : Optional[Any]=None , _lowerCAmelCase : Tuple=None , _lowerCAmelCase : Union[str, Any]=None , **_lowerCAmelCase : Optional[int] ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: SCREAMING_SNAKE_CASE_ = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: SCREAMING_SNAKE_CASE_ = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: SCREAMING_SNAKE_CASE_ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def lowerCAmelCase_ ( self : str , *_lowerCAmelCase : str , **_lowerCAmelCase : Any ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] , *_lowerCAmelCase : Union[str, Any] , **_lowerCAmelCase : str ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def lowerCAmelCase_ ( self : Any ): SCREAMING_SNAKE_CASE_ = self.tokenizer.model_input_names SCREAMING_SNAKE_CASE_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
31
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore SCREAMING_SNAKE_CASE : int = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" SCREAMING_SNAKE_CASE : Dict = [file for file in filepaths if file != file.lower()] if upper_files: print(F"{len(upper_files)} files contain uppercase characters:") print("\n".join(upper_files) + "\n") SCREAMING_SNAKE_CASE : List[Any] = [file for file in filepaths if " " in file] if space_files: print(F"{len(space_files)} files contain space characters:") print("\n".join(space_files) + "\n") SCREAMING_SNAKE_CASE : Any = [file for file in filepaths if "-" in file] if hyphen_files: print(F"{len(hyphen_files)} files contain hyphen characters:") print("\n".join(hyphen_files) + "\n") SCREAMING_SNAKE_CASE : str = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"{len(nodir_files)} files are not in a directory:") print("\n".join(nodir_files) + "\n") SCREAMING_SNAKE_CASE : Tuple = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
89
0
def A__ ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : int ) -> list[str]: """simple docstring""" return [sentence[i : i + ngram_size] for i in range(len(SCREAMING_SNAKE_CASE_ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
32
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Any: _lowercase : str = 10 _lowercase : List[str] = datasets.Features( { 'tokens': datasets.Sequence(datasets.Value('string' ) ), 'labels': datasets.Sequence(datasets.ClassLabel(names=['negative', 'positive'] ) ), 'answers': datasets.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), 'id': datasets.Value('int64' ), } ) _lowercase : Union[str, Any] = datasets.Dataset.from_dict( { 'tokens': [['foo'] * 5] * n, 'labels': [[1] * 5] * n, 'answers': [{'answer_start': [97], 'text': ['1976']}] * 10, 'id': list(range(lowerCamelCase_ ) ), } , features=lowerCamelCase_ , ) return dataset @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : int = str(tmp_path_factory.mktemp('data' ) / 'file.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return filename # FILE_CONTENT + files SCREAMING_SNAKE_CASE : str = "\\n Text data.\n Second line of data." @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'file.txt' _lowercase : List[str] = FILE_CONTENT with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: import bza _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.bz2' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'file.txt.gz' ) _lowercase : Optional[int] = bytes(lowerCamelCase_ , 'utf-8' ) with gzip.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: if datasets.config.LZ4_AVAILABLE: import lza.frame _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.lz4' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with lza.frame.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: if datasets.config.PY7ZR_AVAILABLE: import pyazr _lowercase : int = tmp_path_factory.mktemp('data' ) / 'file.txt.7z' with pyazr.SevenZipFile(lowerCamelCase_ , 'w' ) as archive: archive.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: import tarfile _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: import lzma _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.xz' _lowercase : int = bytes(lowerCamelCase_ , 'utf-8' ) with lzma.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: import zipfile _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'file.txt.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.zst' _lowercase : Dict = bytes(lowerCamelCase_ , 'utf-8' ) with zstd.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'file.xml' _lowercase : Optional[Any] = textwrap.dedent( '\\n <?xml version="1.0" encoding="UTF-8" ?>\n <tmx version="1.4">\n <header segtype="sentence" srclang="ca" />\n <body>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang="en"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang="en"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang="en"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang="en"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang="en"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>' ) with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] SCREAMING_SNAKE_CASE : Optional[Any] = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } SCREAMING_SNAKE_CASE : Tuple = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] SCREAMING_SNAKE_CASE : Any = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> List[str]: return DATA_DICT_OF_LISTS @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Optional[int] = datasets.Dataset.from_dict(lowerCamelCase_ ) _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.sqlite' ) with contextlib.closing(sqlitea.connect(lowerCamelCase_ ) ) as con: _lowercase : Union[str, Any] = con.cursor() cur.execute('CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)' ) for item in DATA: cur.execute('INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : Tuple = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset2.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : str = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: import bza _lowercase : int = tmp_path_factory.mktemp('data' ) / 'dataset.csv.bz2' with open(lowerCamelCase_ , 'rb' ) as f: _lowercase : int = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(csv_path.replace('.csv' , '.CSV' ) ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(csva_path.replace('.csv' , '.CSV' ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.parquet' ) _lowercase : Optional[Any] = pa.schema( { 'col_1': pa.string(), 'col_2': pa.intaa(), 'col_3': pa.floataa(), } ) with open(lowerCamelCase_ , 'wb' ) as f: _lowercase : List[str] = pq.ParquetWriter(lowerCamelCase_ , schema=lowerCamelCase_ ) _lowercase : Any = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(lowerCamelCase_ ) )] for k in DATA[0]} , schema=lowerCamelCase_ ) writer.write_table(lowerCamelCase_ ) writer.close() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : List[Any] = {'data': DATA} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : Optional[Any] = {'data': DATA_DICT_OF_LISTS} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Dict = str(tmp_path_factory.mktemp('data' ) / 'dataset2.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset_312.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_312: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset-str.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_STR: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: import gzip _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Dict: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Tuple = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Optional[int] = ['0', '1', '2', '3'] _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : str = ['0', '1', '2', '3'] _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : List[Any] = ['0', '1', '2', '3'] _lowercase : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset.abc' with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: _lowercase : List[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.ext.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported.ext' ) ) f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported_2.ext' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = '\n'.join(['First', 'Second\u2029with Unicode new line', 'Third'] ) _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset_with_unicode_new_lines.txt' ) with open(lowerCamelCase_ , 'w' , encoding='utf-8' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Dict: return os.path.join('tests' , 'features' , 'data' , 'test_image_rgb.jpg' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> int: return os.path.join('tests' , 'features' , 'data' , 'test_audio_44100.wav' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'dataset.img.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ).replace('.jpg' , '2.jpg' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data_dir' ) (data_dir / "subdir").mkdir() with open(data_dir / 'subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / 'subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden file with open(data_dir / 'subdir' / '.test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '.subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / '.subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) return data_dir
89
0
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Tuple = ['vqvae'] def __init__( self:Any , _a:AutoencoderKL , _a:UNetaDConditionModel , _a:Mel , _a:Union[DDIMScheduler, DDPMScheduler] , ): super().__init__() self.register_modules(unet=_a , scheduler=_a , mel=_a , vqvae=_a ) def SCREAMING_SNAKE_CASE__ ( self:int ): return 50 if isinstance(self.scheduler , _a ) else 10_00 @torch.no_grad() def __call__( self:Tuple , _a:int = 1 , _a:str = None , _a:np.ndarray = None , _a:int = 0 , _a:int = 0 , _a:int = None , _a:torch.Generator = None , _a:float = 0 , _a:float = 0 , _a:torch.Generator = None , _a:float = 0 , _a:torch.Tensor = None , _a:torch.Tensor = None , _a:Union[str, Any]=True , ): snake_case__ = steps or self.get_default_steps() self.scheduler.set_timesteps(_a ) snake_case__ = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: snake_case__ = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: snake_case__ = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=_a , device=self.device , ) snake_case__ = noise snake_case__ = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(_a , _a ) snake_case__ = self.mel.audio_slice_to_image(_a ) snake_case__ = np.frombuffer(input_image.tobytes() , dtype='''uint8''' ).reshape( (input_image.height, input_image.width) ) snake_case__ = (input_image / 2_55) * 2 - 1 snake_case__ = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: snake_case__ = self.vqvae.encode(torch.unsqueeze(_a , 0 ) ).latent_dist.sample( generator=_a )[0] snake_case__ = self.vqvae.config.scaling_factor * input_images if start_step > 0: snake_case__ = self.scheduler.add_noise(_a , _a , self.scheduler.timesteps[start_step - 1] ) snake_case__ = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) snake_case__ = int(mask_start_secs * pixels_per_second ) snake_case__ = int(mask_end_secs * pixels_per_second ) snake_case__ = self.scheduler.add_noise(_a , _a , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , _a ): snake_case__ = self.unet(_a , _a , _a )['''sample'''] else: snake_case__ = self.unet(_a , _a )['''sample'''] if isinstance(self.scheduler , _a ): snake_case__ = self.scheduler.step( model_output=_a , timestep=_a , sample=_a , eta=_a , generator=_a , )['''prev_sample'''] else: snake_case__ = self.scheduler.step( model_output=_a , timestep=_a , sample=_a , generator=_a , )['''prev_sample'''] if mask is not None: if mask_start > 0: snake_case__ = mask[:, step, :, :mask_start] if mask_end > 0: snake_case__ = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance snake_case__ = 1 / self.vqvae.config.scaling_factor * images snake_case__ = self.vqvae.decode(_a )['''sample'''] snake_case__ = (images / 2 + 0.5).clamp(0 , 1 ) snake_case__ = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() snake_case__ = (images * 2_55).round().astype('''uint8''' ) snake_case__ = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(_a , mode='''RGB''' ).convert('''L''' ) for _ in images) ) snake_case__ = [self.mel.image_to_audio(_a ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(_a )[:, np.newaxis, :] ) , **ImagePipelineOutput(_a ) ) @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( self:List[str] , _a:List[Image.Image] , _a:int = 50 ): assert isinstance(self.scheduler , _a ) self.scheduler.set_timesteps(_a ) snake_case__ = np.array( [np.frombuffer(image.tobytes() , dtype='''uint8''' ).reshape((1, image.height, image.width) ) for image in images] ) snake_case__ = (sample / 2_55) * 2 - 1 snake_case__ = torch.Tensor(_a ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): snake_case__ = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps snake_case__ = self.scheduler.alphas_cumprod[t] snake_case__ = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) snake_case__ = 1 - alpha_prod_t snake_case__ = self.unet(_a , _a )['''sample'''] snake_case__ = (1 - alpha_prod_t_prev) ** 0.5 * model_output snake_case__ = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) snake_case__ = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:torch.Tensor , _a:torch.Tensor , _a:float ): snake_case__ = acos(torch.dot(torch.flatten(_a ) , torch.flatten(_a ) ) / torch.norm(_a ) / torch.norm(_a ) ) return sin((1 - alpha) * theta ) * xa / sin(_a ) + sin(alpha * theta ) * xa / sin(_a )
33
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE : str = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : int = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = '''roformer''' def __init__( self , lowerCamelCase_=5_0_0_0_0 , lowerCamelCase_=None , lowerCamelCase_=7_6_8 , lowerCamelCase_=1_2 , lowerCamelCase_=1_2 , lowerCamelCase_=3_0_7_2 , lowerCamelCase_="gelu" , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=1_5_3_6 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=1e-12 , lowerCamelCase_=0 , lowerCamelCase_=False , lowerCamelCase_=True , **lowerCamelCase_ , ) -> Optional[Any]: super().__init__(pad_token_id=lowerCamelCase_ , **lowerCamelCase_) UpperCamelCase = vocab_size UpperCamelCase = hidden_size if embedding_size is None else embedding_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = hidden_act UpperCamelCase = intermediate_size UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = type_vocab_size UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = rotary_value UpperCamelCase = use_cache class snake_case_ ( lowerCamelCase_ ): """simple docstring""" @property def UpperCAmelCase__ ( self) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCamelCase = {0: '''batch''', 1: '''sequence'''} UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ])
34
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=7, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=2, lowerCamelCase=99, lowerCamelCase=0, lowerCamelCase=32, lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=0.1, lowerCamelCase=0.1, lowerCamelCase=5_12, lowerCamelCase=2, lowerCamelCase=0.0_2, lowerCamelCase=2, lowerCamelCase=4, lowerCamelCase="last", lowerCamelCase=True, lowerCamelCase=None, lowerCamelCase=0, ) -> str: """simple docstring""" _lowercase : Union[str, Any] = parent _lowercase : Optional[Any] = batch_size _lowercase : List[str] = seq_length _lowercase : int = is_training _lowercase : List[str] = use_input_lengths _lowercase : int = use_token_type_ids _lowercase : Any = use_labels _lowercase : Union[str, Any] = gelu_activation _lowercase : List[str] = sinusoidal_embeddings _lowercase : str = causal _lowercase : Optional[int] = asm _lowercase : Union[str, Any] = n_langs _lowercase : List[Any] = vocab_size _lowercase : Any = n_special _lowercase : Any = hidden_size _lowercase : str = num_hidden_layers _lowercase : Union[str, Any] = num_attention_heads _lowercase : Tuple = hidden_dropout_prob _lowercase : Optional[int] = attention_probs_dropout_prob _lowercase : Union[str, Any] = max_position_embeddings _lowercase : List[str] = type_sequence_label_size _lowercase : Any = initializer_range _lowercase : int = num_labels _lowercase : Optional[int] = num_choices _lowercase : Optional[Any] = summary_type _lowercase : Optional[Any] = use_proj _lowercase : int = scope _lowercase : List[Any] = bos_token_id def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Tuple = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) _lowercase : List[str] = random_attention_mask([self.batch_size, self.seq_length]) _lowercase : int = None if self.use_input_lengths: _lowercase : Dict = ( ids_tensor([self.batch_size], vocab_size=2) + self.seq_length - 2 ) # small variation of seq_length _lowercase : Tuple = None if self.use_token_type_ids: _lowercase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length], self.n_langs) _lowercase : Tuple = None _lowercase : int = None _lowercase : int = None if self.use_labels: _lowercase : str = ids_tensor([self.batch_size], self.type_sequence_label_size) _lowercase : str = ids_tensor([self.batch_size, self.seq_length], self.num_labels) _lowercase : Dict = ids_tensor([self.batch_size], 2).float() _lowercase : Tuple = ids_tensor([self.batch_size], self.num_choices) _lowercase : Dict = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return XLMConfig( vocab_size=self.vocab_size, n_special=self.n_special, emb_dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, gelu_activation=self.gelu_activation, sinusoidal_embeddings=self.sinusoidal_embeddings, asm=self.asm, causal=self.causal, n_langs=self.n_langs, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, summary_type=self.summary_type, use_proj=self.use_proj, num_labels=self.num_labels, bos_token_id=self.bos_token_id, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Tuple: """simple docstring""" _lowercase : List[Any] = XLMModel(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = model(lowerCamelCase, lengths=lowerCamelCase, langs=lowerCamelCase) _lowercase : int = model(lowerCamelCase, langs=lowerCamelCase) _lowercase : Any = model(lowerCamelCase) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[Any]: """simple docstring""" _lowercase : Dict = XLMWithLMHeadModel(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> str: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnsweringSimple(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Dict = model(lowerCamelCase) _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) _lowercase : Any = outputs self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnswering(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = model(lowerCamelCase) _lowercase : List[Any] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, p_mask=lowerCamelCase, ) _lowercase : List[str] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, ) ((_lowercase) , ) : Optional[Any] = result_with_labels.to_tuple() _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) ((_lowercase) , ) : Any = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape, ()) self.parent.assertEqual(result.start_top_log_probs.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual(result.start_top_index.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual( result.end_top_log_probs.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual( result.end_top_index.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual(result.cls_logits.shape, (self.batch_size,)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> int: """simple docstring""" _lowercase : Optional[Any] = XLMForSequenceClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[int] = model(lowerCamelCase) _lowercase : Optional[int] = model(lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[str]: """simple docstring""" _lowercase : Any = self.num_labels _lowercase : str = XLMForTokenClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, attention_mask=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Dict: """simple docstring""" _lowercase : Optional[Any] = self.num_choices _lowercase : Optional[int] = XLMForMultipleChoice(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = input_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : int = token_type_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : Optional[Any] = input_mask.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : List[str] = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices)) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Dict = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[Any] = config_and_inputs _lowercase : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return config, inputs_dict @require_torch class _lowerCamelCase( _a, _a, _a, unittest.TestCase ): lowercase_ : Any = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) lowercase_ : Optional[int] = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable lowercase_ : Union[str, Any] = ( { """feature-extraction""": XLMModel, """fill-mask""": XLMWithLMHeadModel, """question-answering""": XLMForQuestionAnsweringSimple, """text-classification""": XLMForSequenceClassification, """text-generation""": XLMWithLMHeadModel, """token-classification""": XLMForTokenClassification, """zero-shot""": XLMForSequenceClassification, } if is_torch_available() else {} ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast') ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase=False) -> Optional[int]: """simple docstring""" _lowercase : Any = super()._prepare_for_class(lowerCamelCase, lowerCamelCase, return_labels=lowerCamelCase) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": _lowercase : Any = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) _lowercase : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) return inputs_dict def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Union[str, Any] = XLMModelTester(self) _lowercase : List[str] = ConfigTester(self, config_class=lowerCamelCase, emb_dim=37) def UpperCamelCase ( self) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*lowerCamelCase) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*lowerCamelCase) def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> int: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_attentions in attentions], [True] * len(lowerCamelCase)) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_attentions in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : Dict = min_length + idx + 1 _lowercase : int = min_length + idx + 1 _lowercase : Dict = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions], [expected_shape] * len(lowerCamelCase)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> List[Any]: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_hidden_states in hidden_states], [True] * len(lowerCamelCase), ) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_hidden_states in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : int = min_length + idx + 1 _lowercase : int = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states], [expected_shape] * len(lowerCamelCase), ) pass @slow def UpperCamelCase ( self) -> int: """simple docstring""" for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Dict = XLMModel.from_pretrained(lowerCamelCase) self.assertIsNotNone(lowerCamelCase) @require_torch class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMWithLMHeadModel.from_pretrained('xlm-mlm-en-2048') model.to(lowerCamelCase) _lowercase : Union[str, Any] = torch.tensor([[14, 4_47]], dtype=torch.long, device=lowerCamelCase) # the president _lowercase : Any = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference _lowercase : str = model.generate(lowerCamelCase, do_sample=lowerCamelCase) self.assertListEqual(output_ids[0].cpu().numpy().tolist(), lowerCamelCase)
89
0
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL a_ :str = logging.get_logger(__name__) def a ( A__ , A__ , A__ , A__ ) -> Tuple[int, int]: '''simple docstring''' def constraint_to_multiple_of(A__ , A__ , A__=0 , A__=None ): SCREAMING_SNAKE_CASE__ : Optional[int] = round(val / multiple ) * multiple if max_val is not None and x > max_val: SCREAMING_SNAKE_CASE__ : Any = math.floor(val / multiple ) * multiple if x < min_val: SCREAMING_SNAKE_CASE__ : Any = math.ceil(val / multiple ) * multiple return x SCREAMING_SNAKE_CASE__ : Union[str, Any] = (output_size, output_size) if isinstance(A__ , A__ ) else output_size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Tuple = get_image_size(A__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = output_size # determine new height and width SCREAMING_SNAKE_CASE__ : List[str] = output_height / input_height SCREAMING_SNAKE_CASE__ : Dict = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width SCREAMING_SNAKE_CASE__ : List[str] = scale_width else: # fit height SCREAMING_SNAKE_CASE__ : Optional[Any] = scale_height SCREAMING_SNAKE_CASE__ : int = constraint_to_multiple_of(scale_height * input_height , multiple=A__ ) SCREAMING_SNAKE_CASE__ : int = constraint_to_multiple_of(scale_width * input_width , multiple=A__ ) return (new_height, new_width) class lowercase ( _UpperCAmelCase ): lowerCamelCase : List[str] = ['''pixel_values'''] def __init__( self : List[Any] , _lowercase : bool = True , _lowercase : Dict[str, int] = None , _lowercase : PILImageResampling = PILImageResampling.BILINEAR , _lowercase : bool = False , _lowercase : int = 1 , _lowercase : bool = True , _lowercase : Union[int, float] = 1 / 2_55 , _lowercase : bool = True , _lowercase : Optional[Union[float, List[float]]] = None , _lowercase : Optional[Union[float, List[float]]] = None , **_lowercase : List[Any] , ): super().__init__(**_lowercase ) SCREAMING_SNAKE_CASE__ : Dict = size if size is not None else {'''height''': 3_84, '''width''': 3_84} SCREAMING_SNAKE_CASE__ : Optional[int] = get_size_dict(_lowercase ) SCREAMING_SNAKE_CASE__ : List[str] = do_resize SCREAMING_SNAKE_CASE__ : Optional[int] = size SCREAMING_SNAKE_CASE__ : int = keep_aspect_ratio SCREAMING_SNAKE_CASE__ : Optional[Any] = ensure_multiple_of SCREAMING_SNAKE_CASE__ : List[str] = resample SCREAMING_SNAKE_CASE__ : Union[str, Any] = do_rescale SCREAMING_SNAKE_CASE__ : Optional[int] = rescale_factor SCREAMING_SNAKE_CASE__ : List[Any] = do_normalize SCREAMING_SNAKE_CASE__ : Tuple = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE__ : Any = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase__ ( self : Optional[int] , _lowercase : np.ndarray , _lowercase : Dict[str, int] , _lowercase : bool = False , _lowercase : int = 1 , _lowercase : PILImageResampling = PILImageResampling.BICUBIC , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Optional[int] , ): SCREAMING_SNAKE_CASE__ : List[Any] = get_size_dict(_lowercase ) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_resize_output_image_size( _lowercase , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=_lowercase , multiple=_lowercase , ) return resize(_lowercase , size=_lowercase , resample=_lowercase , data_format=_lowercase , **_lowercase ) def lowercase__ ( self : List[str] , _lowercase : np.ndarray , _lowercase : Union[int, float] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : str , ): return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def lowercase__ ( self : List[str] , _lowercase : np.ndarray , _lowercase : Union[float, List[float]] , _lowercase : Union[float, List[float]] , _lowercase : Optional[Union[str, ChannelDimension]] = None , **_lowercase : Optional[Any] , ): return normalize(_lowercase , mean=_lowercase , std=_lowercase , data_format=_lowercase , **_lowercase ) def lowercase__ ( self : Optional[Any] , _lowercase : ImageInput , _lowercase : bool = None , _lowercase : int = None , _lowercase : bool = None , _lowercase : int = None , _lowercase : PILImageResampling = None , _lowercase : bool = None , _lowercase : float = None , _lowercase : bool = None , _lowercase : Optional[Union[float, List[float]]] = None , _lowercase : Optional[Union[float, List[float]]] = None , _lowercase : Optional[Union[str, TensorType]] = None , _lowercase : ChannelDimension = ChannelDimension.FIRST , **_lowercase : Tuple , ): SCREAMING_SNAKE_CASE__ : Union[str, Any] = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE__ : List[Any] = size if size is not None else self.size SCREAMING_SNAKE_CASE__ : List[str] = get_size_dict(_lowercase ) SCREAMING_SNAKE_CASE__ : List[Any] = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio SCREAMING_SNAKE_CASE__ : List[str] = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of SCREAMING_SNAKE_CASE__ : Tuple = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE__ : str = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE__ : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE__ : Optional[int] = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE__ : Tuple = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE__ : str = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE__ : Optional[Any] = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE__ : str = [to_numpy_array(_lowercase ) for image in images] if do_resize: SCREAMING_SNAKE_CASE__ : Any = [self.resize(image=_lowercase , size=_lowercase , resample=_lowercase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE__ : Tuple = [self.rescale(image=_lowercase , scale=_lowercase ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE__ : Any = [self.normalize(image=_lowercase , mean=_lowercase , std=_lowercase ) for image in images] SCREAMING_SNAKE_CASE__ : Union[str, Any] = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] SCREAMING_SNAKE_CASE__ : str = {'''pixel_values''': images} return BatchFeature(data=_lowercase , tensor_type=_lowercase ) def lowercase__ ( self : Tuple , _lowercase : Optional[Any] , _lowercase : List[Tuple] = None ): SCREAMING_SNAKE_CASE__ : str = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowercase ) != len(_lowercase ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(_lowercase ): SCREAMING_SNAKE_CASE__ : Union[str, Any] = target_sizes.numpy() SCREAMING_SNAKE_CASE__ : Tuple = [] for idx in range(len(_lowercase ) ): SCREAMING_SNAKE_CASE__ : Dict = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=_lowercase ) SCREAMING_SNAKE_CASE__ : Any = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowercase ) else: SCREAMING_SNAKE_CASE__ : Any = logits.argmax(dim=1 ) SCREAMING_SNAKE_CASE__ : Dict = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
35
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.17.0.dev0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/text-classification/requirements.txt") SCREAMING_SNAKE_CASE : Dict = logging.getLogger(__name__) @dataclass class _lowerCamelCase: lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""}, ) lowercase_ : int = field( default=10_24, metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of prediction examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the training data."""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the validation data."""} ) lowercase_ : Optional[str] = field(default=_a, metadata={"""help""": """A csv or a json file containing the test data."""} ) def UpperCamelCase ( self) -> Dict: """simple docstring""" if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError('Need either a GLUE task, a training/validation file or a dataset name.') else: _lowercase : int = self.train_file.split('.')[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." _lowercase : Tuple = self.validation_file.split('.')[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class _lowerCamelCase: lowercase_ : str = field( default=_a, metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""}, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""}, ) lowercase_ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) def UpperCamelCase_( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowercase , _lowercase , _lowercase : Tuple = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowercase , _lowercase , _lowercase : Union[str, Any] = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) _lowercase : Union[str, Any] = training_args.get_process_log_level() logger.setLevel(lowerCamelCase_ ) datasets.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _lowercase : Optional[int] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowercase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. _lowercase : Dict = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. _lowercase : Optional[Any] = {'train': data_args.train_file, 'validation': data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: _lowercase : Tuple = data_args.train_file.split('.' )[-1] _lowercase : int = data_args.test_file.split('.' )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." _lowercase : Any = data_args.test_file else: raise ValueError('Need either a GLUE task or a test file for `do_predict`.' ) for key in data_files.keys(): logger.info(F'''load a local file for {key}: {data_files[key]}''' ) if data_args.train_file.endswith('.csv' ): # Loading a dataset from local csv files _lowercase : str = load_dataset('csv' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files _lowercase : Optional[int] = load_dataset('json' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels _lowercase : Optional[Any] = raw_datasets['train'].features['label'].names _lowercase : Any = len(lowerCamelCase_ ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer _lowercase : str = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=lowerCamelCase_ , ) _lowercase : Tuple = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: _lowercase : int = 'max_length' else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _lowercase : str = False # Some models have set the order of the labels to use, so let's make sure we do use it. _lowercase : List[Any] = {'Refused': 0, 'Entailed': 1} _lowercase : Union[str, Any] = {0: 'Refused', 1: 'Entailed'} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) _lowercase : List[str] = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(lowerCamelCase_ ): # Tokenize the texts def _convert_table_text_to_pandas(lowerCamelCase_ ): _lowercase : int = [_table_row.split('#' ) for _table_row in _table_text.strip('\n' ).split('\n' )] _lowercase : Any = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd _lowercase : List[Any] = examples['statement'] _lowercase : Optional[Any] = list(map(_convert_table_text_to_pandas , examples['table_text'] ) ) _lowercase : Union[str, Any] = tokenizer(lowerCamelCase_ , lowerCamelCase_ , padding=lowerCamelCase_ , max_length=lowerCamelCase_ , truncation=lowerCamelCase_ ) _lowercase : Any = examples['label'] return result with training_args.main_process_first(desc='dataset map pre-processing' ): _lowercase : str = raw_datasets.map( lowerCamelCase_ , batched=lowerCamelCase_ , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on dataset' , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _lowercase : Any = raw_datasets['train'] if data_args.max_train_samples is not None: _lowercase : str = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _lowercase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _lowercase : List[Any] = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError('--do_predict requires a test dataset' ) _lowercase : Optional[int] = raw_datasets['test'] if data_args.max_predict_samples is not None: _lowercase : List[str] = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(lowerCamelCase_ ) ) , 3 ): logger.info(F'''Sample {index} of the training set: {train_dataset[index]}.''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCamelCase_ ): _lowercase : Dict = p.predictions[0] if isinstance(p.predictions , lowerCamelCase_ ) else p.predictions _lowercase : Tuple = np.argmax(lowerCamelCase_ , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _lowercase : Any = default_data_collator elif training_args.fpaa: _lowercase : str = DataCollatorWithPadding(lowerCamelCase_ , pad_to_multiple_of=8 ) else: _lowercase : Optional[Any] = None # Initialize our Trainer _lowercase : List[str] = Trainer( model=lowerCamelCase_ , args=lowerCamelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=lowerCamelCase_ , tokenizer=lowerCamelCase_ , data_collator=lowerCamelCase_ , ) # Training if training_args.do_train: _lowercase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _lowercase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowercase : Optional[Any] = last_checkpoint _lowercase : Optional[Any] = trainer.train(resume_from_checkpoint=lowerCamelCase_ ) _lowercase : List[Any] = train_result.metrics _lowercase : Dict = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCamelCase_ ) ) _lowercase : int = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics('train' , lowerCamelCase_ ) trainer.save_metrics('train' , lowerCamelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _lowercase : Tuple = trainer.evaluate(eval_dataset=lowerCamelCase_ ) _lowercase : Any = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCamelCase_ ) _lowercase : Optional[int] = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.log_metrics('eval' , lowerCamelCase_ ) trainer.save_metrics('eval' , lowerCamelCase_ ) if training_args.do_predict: logger.info('*** Predict ***' ) # Removing the `label` columns because it contains -1 and Trainer won't like that. _lowercase : Any = predict_dataset.remove_columns('label' ) _lowercase : Optional[Any] = trainer.predict(lowerCamelCase_ , metric_key_prefix='predict' ).predictions _lowercase : Union[str, Any] = np.argmax(lowerCamelCase_ , axis=1 ) _lowercase : Dict = os.path.join(training_args.output_dir , 'predict_results_tabfact.txt' ) if trainer.is_world_process_zero(): with open(lowerCamelCase_ , 'w' ) as writer: logger.info('***** Predict Results *****' ) writer.write('index\tprediction\n' ) for index, item in enumerate(lowerCamelCase_ ): _lowercase : List[str] = label_list[item] writer.write(F'''{index}\t{item}\n''' ) _lowercase : str = {'finetuned_from': model_args.model_name_or_path, 'tasks': 'text-classification'} if training_args.push_to_hub: trainer.push_to_hub(**lowerCamelCase_ ) else: trainer.create_model_card(**lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
89
0
from __future__ import annotations def lowercase ( __A : str , __A : list[str] | None = None ) -> list[list[str]]: '''simple docstring''' snake_case : Optional[int] = word_bank or [] # create a table snake_case : int = len(__A ) + 1 snake_case : list[list[list[str]]] = [] for _ in range(__A ): table.append([] ) # seed value snake_case : Optional[Any] = [[]] # because empty string has empty combination # iterate through the indices for i in range(__A ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(__A )] == word: snake_case : list[list[str]] = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(__A )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(__A )]: combination.reverse() return table[len(__A )] if __name__ == "__main__": print(all_construct('''jwajalapa''', ['''jwa''', '''j''', '''w''', '''a''', '''la''', '''lapa'''])) print(all_construct('''rajamati''', ['''s''', '''raj''', '''amat''', '''raja''', '''ma''', '''i''', '''t'''])) print( all_construct( '''hexagonosaurus''', ['''h''', '''ex''', '''hex''', '''ag''', '''ago''', '''ru''', '''auru''', '''rus''', '''go''', '''no''', '''o''', '''s'''], ) )
36
from maths.prime_factors import prime_factors def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): _lowercase : str = F'''Input value of [number={number}] must be an integer''' raise TypeError(lowerCamelCase_ ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(lowerCamelCase_ ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
89
0
import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase : Tuple = logging.get_logger(__name__) UpperCamelCase : str = { """kakaobrain/align-base""": """https://huggingface.co/kakaobrain/align-base/resolve/main/config.json""", } class A__ ( A__ ): """simple docstring""" _lowercase = 'align_text_model' def __init__( self : Dict , lowerCamelCase__ : Dict=30_522 , lowerCamelCase__ : Optional[int]=768 , lowerCamelCase__ : Tuple=12 , lowerCamelCase__ : Tuple=12 , lowerCamelCase__ : List[str]=3_072 , lowerCamelCase__ : Union[str, Any]="gelu" , lowerCamelCase__ : str=0.1 , lowerCamelCase__ : List[Any]=0.1 , lowerCamelCase__ : Any=512 , lowerCamelCase__ : Optional[int]=2 , lowerCamelCase__ : Optional[int]=0.02 , lowerCamelCase__ : Optional[Any]=1E-12 , lowerCamelCase__ : Dict=0 , lowerCamelCase__ : Optional[int]="absolute" , lowerCamelCase__ : List[Any]=True , **lowerCamelCase__ : Optional[Any] , ): super().__init__(**lowerCamelCase__ ) a__ : Optional[Any] = vocab_size a__ : Optional[int] = hidden_size a__ : str = num_hidden_layers a__ : int = num_attention_heads a__ : Any = hidden_act a__ : str = intermediate_size a__ : List[Any] = hidden_dropout_prob a__ : str = attention_probs_dropout_prob a__ : List[Any] = max_position_embeddings a__ : List[Any] = type_vocab_size a__ : Optional[int] = initializer_range a__ : int = layer_norm_eps a__ : Optional[Any] = position_embedding_type a__ : Dict = use_cache a__ : Optional[int] = pad_token_id @classmethod def _UpperCamelCase( cls : int , lowerCamelCase__ : Union[str, os.PathLike] , **lowerCamelCase__ : List[str] ): cls._set_token_in_kwargs(lowerCamelCase__ ) a__, a__ : Dict = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the text config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": a__ : int = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class A__ ( A__ ): """simple docstring""" _lowercase = 'align_vision_model' def __init__( self : List[Any] , lowerCamelCase__ : int = 3 , lowerCamelCase__ : int = 600 , lowerCamelCase__ : float = 2.0 , lowerCamelCase__ : float = 3.1 , lowerCamelCase__ : int = 8 , lowerCamelCase__ : List[int] = [3, 3, 5, 3, 5, 5, 3] , lowerCamelCase__ : List[int] = [32, 16, 24, 40, 80, 112, 192] , lowerCamelCase__ : List[int] = [16, 24, 40, 80, 112, 192, 320] , lowerCamelCase__ : List[int] = [] , lowerCamelCase__ : List[int] = [1, 2, 2, 2, 1, 2, 1] , lowerCamelCase__ : List[int] = [1, 2, 2, 3, 3, 4, 1] , lowerCamelCase__ : List[int] = [1, 6, 6, 6, 6, 6, 6] , lowerCamelCase__ : float = 0.25 , lowerCamelCase__ : str = "swish" , lowerCamelCase__ : int = 2_560 , lowerCamelCase__ : str = "mean" , lowerCamelCase__ : float = 0.02 , lowerCamelCase__ : float = 0.001 , lowerCamelCase__ : float = 0.99 , lowerCamelCase__ : float = 0.2 , **lowerCamelCase__ : List[str] , ): super().__init__(**lowerCamelCase__ ) a__ : Union[str, Any] = num_channels a__ : List[Any] = image_size a__ : int = width_coefficient a__ : int = depth_coefficient a__ : Union[str, Any] = depth_divisor a__ : Optional[Any] = kernel_sizes a__ : Union[str, Any] = in_channels a__ : Optional[Any] = out_channels a__ : Any = depthwise_padding a__ : Any = strides a__ : int = num_block_repeats a__ : List[Any] = expand_ratios a__ : int = squeeze_expansion_ratio a__ : Optional[Any] = hidden_act a__ : List[str] = hidden_dim a__ : List[str] = pooling_type a__ : List[str] = initializer_range a__ : Union[str, Any] = batch_norm_eps a__ : int = batch_norm_momentum a__ : List[str] = drop_connect_rate a__ : Dict = sum(lowerCamelCase__ ) * 4 @classmethod def _UpperCamelCase( cls : Dict , lowerCamelCase__ : Union[str, os.PathLike] , **lowerCamelCase__ : Dict ): cls._set_token_in_kwargs(lowerCamelCase__ ) a__, a__ : Union[str, Any] = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the vision config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": a__ : List[str] = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class A__ ( A__ ): """simple docstring""" _lowercase = 'align' _lowercase = True def __init__( self : str , lowerCamelCase__ : Optional[int]=None , lowerCamelCase__ : Any=None , lowerCamelCase__ : Dict=640 , lowerCamelCase__ : Tuple=1.0 , lowerCamelCase__ : Any=0.02 , **lowerCamelCase__ : Optional[int] , ): super().__init__(**lowerCamelCase__ ) if text_config is None: a__ : Optional[int] = {} logger.info("text_config is None. Initializing the AlignTextConfig with default values." ) if vision_config is None: a__ : Union[str, Any] = {} logger.info("vision_config is None. Initializing the AlignVisionConfig with default values." ) a__ : Optional[Any] = AlignTextConfig(**lowerCamelCase__ ) a__ : str = AlignVisionConfig(**lowerCamelCase__ ) a__ : Dict = projection_dim a__ : List[str] = temperature_init_value a__ : Dict = initializer_range @classmethod def _UpperCamelCase( cls : Tuple , lowerCamelCase__ : AlignTextConfig , lowerCamelCase__ : AlignVisionConfig , **lowerCamelCase__ : List[str] ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowerCamelCase__ ) def _UpperCamelCase( self : Any ): a__ : int = copy.deepcopy(self.__dict__ ) a__ : int = self.text_config.to_dict() a__ : Tuple = self.vision_config.to_dict() a__ : Optional[Any] = self.__class__.model_type return output
37
from __future__ import annotations from typing import Any class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = 0) -> None: """simple docstring""" _lowercase , _lowercase : str = row, column _lowercase : Any = [[default_value for c in range(lowerCamelCase)] for r in range(lowerCamelCase)] def __str__( self) -> str: """simple docstring""" _lowercase : Tuple = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _lowercase : str = 0 for row_vector in self.array: for obj in row_vector: _lowercase : Optional[int] = max(lowerCamelCase, len(str(lowerCamelCase))) _lowercase : List[str] = F'''%{max_element_length}s''' # Make string and return def single_line(lowerCamelCase) -> str: nonlocal string_format_identifier _lowercase : Union[str, Any] = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector) line += "]" return line s += "\n".join(single_line(lowerCamelCase) for row_vector in self.array) return s def __repr__( self) -> str: """simple docstring""" return str(self) def UpperCamelCase ( self, lowerCamelCase) -> bool: """simple docstring""" if not (isinstance(lowerCamelCase, (list, tuple)) and len(lowerCamelCase) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self, lowerCamelCase) -> Any: """simple docstring""" assert self.validate_indicies(lowerCamelCase) return self.array[loc[0]][loc[1]] def __setitem__( self, lowerCamelCase, lowerCamelCase) -> None: """simple docstring""" assert self.validate_indicies(lowerCamelCase) _lowercase : Optional[Any] = value def __add__( self, lowerCamelCase) -> Matrix: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) assert self.row == another.row and self.column == another.column # Add _lowercase : Any = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : int = self[r, c] + another[r, c] return result def __neg__( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : List[str] = -self[r, c] return result def __sub__( self, lowerCamelCase) -> Matrix: """simple docstring""" return self + (-another) def __mul__( self, lowerCamelCase) -> Matrix: """simple docstring""" if isinstance(lowerCamelCase, (int, float)): # Scalar multiplication _lowercase : Dict = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] * another return result elif isinstance(lowerCamelCase, lowerCamelCase): # Matrix multiplication assert self.column == another.row _lowercase : str = Matrix(self.row, another.column) for r in range(self.row): for c in range(another.column): for i in range(self.column): result[r, c] += self[r, i] * another[i, c] return result else: _lowercase : Tuple = F'''Unsupported type given for another ({type(lowerCamelCase)})''' raise TypeError(lowerCamelCase) def UpperCamelCase ( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.column, self.row) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] return result def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> Any: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) and isinstance(lowerCamelCase, lowerCamelCase) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _lowercase : Dict = v.transpose() _lowercase : Any = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def UpperCamelCase_( ) -> None: # a^(-1) _lowercase : Optional[int] = Matrix(3 , 3 , 0 ) for i in range(3 ): _lowercase : int = 1 print(F'''a^(-1) is {ainv}''' ) # u, v _lowercase : Dict = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : Dict = 1, 2, -3 _lowercase : List[Any] = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : int = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCamelCase_ , lowerCamelCase_ )}''' ) def UpperCamelCase_( ) -> None: import doctest doctest.testmod() testa()
89
0
'''simple docstring''' def UpperCamelCase__ ( __magic_name__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: '''simple docstring''' snake_case__ : Tuple = set() # Replace all the whitespace in our sentence snake_case__ : List[Any] = input_str.replace(""" """ , """""" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(__magic_name__ ) == 26 def UpperCamelCase__ ( __magic_name__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: '''simple docstring''' snake_case__ : Optional[Any] = [False] * 26 for char in input_str: if char.islower(): snake_case__ : int = True elif char.isupper(): snake_case__ : Optional[Any] = True return all(__magic_name__ ) def UpperCamelCase__ ( __magic_name__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: '''simple docstring''' return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def UpperCamelCase__ ( ) -> None: '''simple docstring''' from timeit import timeit snake_case__ : Optional[Any] = """from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest""" print(timeit("""is_pangram()""" , setup=__magic_name__ ) ) print(timeit("""is_pangram_faster()""" , setup=__magic_name__ ) ) print(timeit("""is_pangram_fastest()""" , setup=__magic_name__ ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
38
import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = int(lowerCamelCase_ ) _lowercase , _lowercase , _lowercase : Optional[Any] = t // 3600, (t // 60) % 60, t % 60 return F'''{h}:{m:02d}:{s:02d}''' if h != 0 else F'''{m:02d}:{s:02d}''' def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=300 ) -> Dict: # docstyle-ignore return F''' <div> {prefix} <progress value=\'{value}\' max=\'{total}\' style=\'width:{width}px; height:20px; vertical-align: middle;\'></progress> {label} </div> ''' def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : int = '<table border="1" class="dataframe">\n' html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += F''' <th>{i}</th>\n''' html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: _lowercase : Any = F'''{elt:.6f}''' if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else str(lowerCamelCase_ ) html_code += F''' <td>{elt}</td>\n''' html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class _lowerCamelCase: lowercase_ : str = 5 lowercase_ : str = 0.2 def __init__( self, lowerCamelCase, lowerCamelCase = None, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = 3_00, ) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = total _lowercase : Optional[int] = '' if prefix is None else prefix _lowercase : Tuple = leave _lowercase : str = parent _lowercase : str = width _lowercase : List[Any] = None _lowercase : List[str] = None _lowercase : Tuple = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, lowerCamelCase = None) -> Dict: """simple docstring""" _lowercase : Any = value if comment is not None: _lowercase : Union[str, Any] = comment if self.last_value is None: _lowercase : Dict = time.time() _lowercase : Tuple = value _lowercase : str = None _lowercase : Optional[int] = self.warmup _lowercase : Optional[Any] = 1 self.update_bar(lowerCamelCase) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for, self.total): if self.first_calls > 0: self.first_calls -= 1 _lowercase : List[str] = time.time() _lowercase : Tuple = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: _lowercase : Dict = self.elapsed_time / (value - self.start_value) else: _lowercase : int = None if value >= self.total: _lowercase : Dict = self.total _lowercase : List[str] = None if not self.leave: self.close() elif self.average_time_per_item is not None: _lowercase : Optional[int] = self.average_time_per_item * (self.total - value) self.update_bar(lowerCamelCase) _lowercase : int = value _lowercase : Tuple = current_time if self.average_time_per_item is None: _lowercase : str = 1 else: _lowercase : int = max(int(self.update_every / self.average_time_per_item), 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None) -> Optional[Any]: """simple docstring""" _lowercase : List[Any] = ' ' * (len(str(self.total)) - len(str(lowerCamelCase))) + str(lowerCamelCase) if self.elapsed_time is None: _lowercase : int = F'''[{spaced_value}/{self.total} : < :''' elif self.predicted_remaining is None: _lowercase : Union[str, Any] = F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)}''' else: _lowercase : Union[str, Any] = ( F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)} <''' F''' {format_time(self.predicted_remaining)}''' ) self.label += F''', {1/self.average_time_per_item:.2f} it/s''' self.label += "]" if self.comment is None or len(self.comment) == 0 else F''', {self.comment}]''' self.display() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: _lowercase : Optional[Any] = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" if self.parent is None and self.output is not None: self.output.update(disp.HTML('')) class _lowerCamelCase( _a ): def __init__( self, lowerCamelCase, lowerCamelCase=None) -> int: """simple docstring""" super().__init__(lowerCamelCase) _lowercase : Optional[Any] = None if column_names is None else [column_names] _lowercase : Any = None def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: _lowercase : Dict = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self, lowerCamelCase) -> Dict: """simple docstring""" if self.inner_table is None: _lowercase : Dict = [list(values.keys()), list(values.values())] else: _lowercase : Tuple = self.inner_table[0] if len(self.inner_table) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(lowerCamelCase) _lowercase : str = columns self.inner_table.append([values[c] for c in columns]) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None, lowerCamelCase=3_00) -> Union[str, Any]: """simple docstring""" _lowercase : List[str] = NotebookProgressBar(lowerCamelCase, prefix=lowerCamelCase, parent=self, width=lowerCamelCase) return self.child_bar def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Optional[Any] = None self.display() class _lowerCamelCase( _a ): def __init__( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = None _lowercase : Dict = None _lowercase : Dict = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Dict: """simple docstring""" _lowercase : Dict = 'Epoch' if args.evaluation_strategy == IntervalStrategy.EPOCH else 'Step' _lowercase : Dict = 0 _lowercase : Tuple = 0 _lowercase : int = [self.first_column] + ['Training Loss'] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append('Validation Loss') _lowercase : Union[str, Any] = NotebookTrainingTracker(state.max_steps, lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" _lowercase : Any = int(state.epoch) if int(state.epoch) == state.epoch else F'''{state.epoch:.2f}''' self.training_tracker.update( state.global_step + 1, comment=F'''Epoch {epoch}/{state.num_train_epochs}''', force_update=self._force_next_update, ) _lowercase : str = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Any: """simple docstring""" if not has_length(lowerCamelCase): return if self.prediction_bar is None: if self.training_tracker is not None: _lowercase : Optional[int] = self.training_tracker.add_child(len(lowerCamelCase)) else: _lowercase : Optional[int] = NotebookProgressBar(len(lowerCamelCase)) self.prediction_bar.update(1) else: self.prediction_bar.update(self.prediction_bar.value + 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Optional[int]: """simple docstring""" if self.prediction_bar is not None: self.prediction_bar.close() _lowercase : Any = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[Any]: """simple docstring""" if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: _lowercase : Dict = {'Training Loss': logs['loss']} # First column is necessarily Step sine we're not in epoch eval strategy _lowercase : List[Any] = state.global_step self.training_tracker.write_line(lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[str]: """simple docstring""" if self.training_tracker is not None: _lowercase : Tuple = {'Training Loss': 'No log', 'Validation Loss': 'No log'} for log in reversed(state.log_history): if "loss" in log: _lowercase : int = log['loss'] break if self.first_column == "Epoch": _lowercase : Union[str, Any] = int(state.epoch) else: _lowercase : Optional[Any] = state.global_step _lowercase : str = 'eval' for k in metrics: if k.endswith('_loss'): _lowercase : str = re.sub(R'\_loss$', '', lowerCamelCase) _lowercase : Tuple = metrics.pop('total_flos', lowerCamelCase) _lowercase : List[str] = metrics.pop('epoch', lowerCamelCase) _lowercase : List[Any] = metrics.pop(F'''{metric_key_prefix}_runtime''', lowerCamelCase) _lowercase : Dict = metrics.pop(F'''{metric_key_prefix}_samples_per_second''', lowerCamelCase) _lowercase : Tuple = metrics.pop(F'''{metric_key_prefix}_steps_per_second''', lowerCamelCase) _lowercase : List[str] = metrics.pop(F'''{metric_key_prefix}_jit_compilation_time''', lowerCamelCase) for k, v in metrics.items(): if k == F'''{metric_key_prefix}_loss''': _lowercase : Union[str, Any] = v else: _lowercase : Optional[Any] = k.split('_') _lowercase : Optional[int] = ' '.join([part.capitalize() for part in splits[1:]]) _lowercase : Tuple = v self.training_tracker.write_line(lowerCamelCase) self.training_tracker.remove_child() _lowercase : str = None # Evaluation takes a long time so we should force the next update. _lowercase : Optional[Any] = True def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" self.training_tracker.update( state.global_step, comment=F'''Epoch {int(state.epoch)}/{state.num_train_epochs}''', force_update=lowerCamelCase) _lowercase : Any = None
89
0
from __future__ import annotations def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): # noqa: E741 while r - l > 1: snake_case_ = (l + r) // 2 if v[m] >= key: snake_case_ = m else: snake_case_ = m # noqa: E741 return r def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): if len(SCREAMING_SNAKE_CASE__ ) == 0: return 0 snake_case_ = [0] * len(SCREAMING_SNAKE_CASE__ ) snake_case_ = 1 snake_case_ = v[0] for i in range(1 , len(SCREAMING_SNAKE_CASE__ ) ): if v[i] < tail[0]: snake_case_ = v[i] elif v[i] > tail[length - 1]: snake_case_ = v[i] length += 1 else: snake_case_ = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
39
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 18, 2] _lowercase : Tuple = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Any = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Dict = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: _lowercase : Any = [3, 3, 3, 3] _lowercase : Any = [5, 5, 5, 5] elif "fl4" in model_name: _lowercase : Dict = [4, 4, 4, 4] _lowercase : Tuple = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: _lowercase : str = [3, 3, 3, 3] if "lrf" in model_name: _lowercase : Optional[int] = [3, 3, 3, 3] else: _lowercase : Dict = [2, 2, 2, 2] if "tiny" in model_name: _lowercase : List[str] = 96 elif "small" in model_name: _lowercase : Dict = 96 elif "base" in model_name: _lowercase : Optional[int] = 128 elif "large" in model_name: _lowercase : List[Any] = 192 elif "xlarge" in model_name: _lowercase : Optional[Any] = 256 elif "huge" in model_name: _lowercase : Dict = 352 # set label information _lowercase : int = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: _lowercase : str = 'imagenet-22k-id2label.json' else: _lowercase : Tuple = 'imagenet-1k-id2label.json' _lowercase : Union[str, Any] = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : Any = {v: k for k, v in idalabel.items()} _lowercase : Optional[Any] = FocalNetConfig( embed_dim=lowerCamelCase_ , depths=lowerCamelCase_ , focal_levels=lowerCamelCase_ , focal_windows=lowerCamelCase_ , use_conv_embed=lowerCamelCase_ , idalabel=lowerCamelCase_ , labelaid=lowerCamelCase_ , use_post_layernorm=lowerCamelCase_ , use_layerscale=lowerCamelCase_ , ) return config def UpperCamelCase_( lowerCamelCase_ ) -> Any: if "patch_embed.proj" in name: _lowercase : Optional[Any] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: _lowercase : str = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: _lowercase : Any = 'encoder.' + name if "encoder.layers" in name: _lowercase : int = name.replace('encoder.layers' , 'encoder.stages' ) if "downsample.proj" in name: _lowercase : Tuple = name.replace('downsample.proj' , 'downsample.projection' ) if "blocks" in name: _lowercase : str = name.replace('blocks' , 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: _lowercase : List[str] = name.replace('modulation.f' , 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: _lowercase : int = name.replace('modulation.h' , 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: _lowercase : Any = name.replace('modulation.proj' , 'modulation.projection_out' ) if name == "norm.weight": _lowercase : Any = 'layernorm.weight' if name == "norm.bias": _lowercase : Tuple = 'layernorm.bias' if "head" in name: _lowercase : Optional[int] = name.replace('head' , 'classifier' ) else: _lowercase : Optional[int] = 'focalnet.' + name return name def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False ) -> str: # fmt: off _lowercase : Dict = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on _lowercase : Dict = model_name_to_url[model_name] print('Checkpoint URL: ' , lowerCamelCase_ ) _lowercase : List[str] = torch.hub.load_state_dict_from_url(lowerCamelCase_ , map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[int] = val _lowercase : Union[str, Any] = get_focalnet_config(lowerCamelCase_ ) _lowercase : Optional[Any] = FocalNetForImageClassification(lowerCamelCase_ ) model.eval() # load state dict model.load_state_dict(lowerCamelCase_ ) # verify conversion _lowercase : Optional[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : Any = BitImageProcessor( do_resize=lowerCamelCase_ , size={'shortest_edge': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=lowerCamelCase_ , crop_size=224 , do_normalize=lowerCamelCase_ , image_mean=lowerCamelCase_ , image_std=lowerCamelCase_ , ) _lowercase : List[str] = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) _lowercase : List[Any] = processor(images=lowerCamelCase_ , return_tensors='pt' ) _lowercase : str = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ), ] ) _lowercase : List[str] = image_transforms(lowerCamelCase_ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , lowerCamelCase_ , atol=1e-4 ) _lowercase : Dict = model(**lowerCamelCase_ ) _lowercase : int = outputs.logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) print('First values of logits:' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": _lowercase : Optional[Any] = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": _lowercase : int = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": _lowercase : str = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": _lowercase : Any = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": _lowercase : List[Any] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": _lowercase : int = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3] , lowerCamelCase_ , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_ ) -> Optional[int]: UpperCamelCase : Tuple = n UpperCamelCase : List[Any] = [None] * self.n UpperCamelCase : str = 0 # index of the first element UpperCamelCase : Optional[int] = 0 UpperCamelCase : Dict = 0 def __len__( self ) -> int: return self.size def snake_case_ ( self ) -> bool: return self.size == 0 def snake_case_ ( self ) -> str: return False if self.is_empty() else self.array[self.front] def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> Optional[int]: if self.size >= self.n: raise Exception('QUEUE IS FULL' ) UpperCamelCase : Union[str, Any] = data UpperCamelCase : int = (self.rear + 1) % self.n self.size += 1 return self def snake_case_ ( self ) -> Dict: if self.size == 0: raise Exception('UNDERFLOW' ) UpperCamelCase : Tuple = self.array[self.front] UpperCamelCase : str = None UpperCamelCase : Tuple = (self.front + 1) % self.n self.size -= 1 return temp
40
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Any = { "ut/deta": "https://huggingface.co/ut/deta/resolve/main/config.json", } class _lowerCamelCase( _a ): lowercase_ : Any = """deta""" lowercase_ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self, lowerCamelCase=None, lowerCamelCase=9_00, lowerCamelCase=20_48, lowerCamelCase=6, lowerCamelCase=20_48, lowerCamelCase=8, lowerCamelCase=6, lowerCamelCase=10_24, lowerCamelCase=8, lowerCamelCase=0.0, lowerCamelCase=True, lowerCamelCase="relu", lowerCamelCase=2_56, lowerCamelCase=0.1, lowerCamelCase=0.0, lowerCamelCase=0.0, lowerCamelCase=0.0_2, lowerCamelCase=1.0, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase="sine", lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=4, lowerCamelCase=True, lowerCamelCase=3_00, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=1, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=0.1, lowerCamelCase=0.2_5, **lowerCamelCase, ) -> Any: """simple docstring""" if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.') _lowercase : List[Any] = CONFIG_MAPPING['resnet'](out_features=['stage2', 'stage3', 'stage4']) else: if isinstance(lowerCamelCase, lowerCamelCase): _lowercase : Dict = backbone_config.pop('model_type') _lowercase : int = CONFIG_MAPPING[backbone_model_type] _lowercase : Union[str, Any] = config_class.from_dict(lowerCamelCase) _lowercase : Union[str, Any] = backbone_config _lowercase : Any = num_queries _lowercase : Union[str, Any] = max_position_embeddings _lowercase : Union[str, Any] = d_model _lowercase : Optional[int] = encoder_ffn_dim _lowercase : Optional[int] = encoder_layers _lowercase : Optional[Any] = encoder_attention_heads _lowercase : Optional[Any] = decoder_ffn_dim _lowercase : Dict = decoder_layers _lowercase : Tuple = decoder_attention_heads _lowercase : Union[str, Any] = dropout _lowercase : Optional[Any] = attention_dropout _lowercase : int = activation_dropout _lowercase : Tuple = activation_function _lowercase : List[Any] = init_std _lowercase : Union[str, Any] = init_xavier_std _lowercase : int = encoder_layerdrop _lowercase : Optional[int] = auxiliary_loss _lowercase : Dict = position_embedding_type # deformable attributes _lowercase : Any = num_feature_levels _lowercase : str = encoder_n_points _lowercase : Any = decoder_n_points _lowercase : List[str] = two_stage _lowercase : Dict = two_stage_num_proposals _lowercase : Any = with_box_refine _lowercase : List[Any] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.') # Hungarian matcher _lowercase : List[Any] = class_cost _lowercase : Optional[int] = bbox_cost _lowercase : str = giou_cost # Loss coefficients _lowercase : Optional[int] = mask_loss_coefficient _lowercase : int = dice_loss_coefficient _lowercase : List[Any] = bbox_loss_coefficient _lowercase : Optional[Any] = giou_loss_coefficient _lowercase : str = eos_coefficient _lowercase : int = focal_alpha super().__init__(is_encoder_decoder=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.d_model def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : int = copy.deepcopy(self.__dict__) _lowercase : Optional[int] = self.backbone_config.to_dict() _lowercase : Optional[Any] = self.__class__.model_type return output
89
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ , A__ ): """simple docstring""" __lowercase = b.T __lowercase = np.sum(np.square(A__ ) , axis=1 ) __lowercase = np.sum(np.square(A__ ) , axis=0 ) __lowercase = np.matmul(A__ , A__ ) __lowercase = aa[:, None] - 2 * ab + ba[None, :] return d def _A ( A__ , A__ ): """simple docstring""" __lowercase = x.reshape(-1 , 3 ) __lowercase = squared_euclidean_distance(A__ , A__ ) return np.argmin(A__ , axis=1 ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = ['pixel_values'] def __init__( self : Any ,lowercase__ : Optional[Union[List[List[int]], np.ndarray]] = None ,lowercase__ : bool = True ,lowercase__ : Dict[str, int] = None ,lowercase__ : PILImageResampling = PILImageResampling.BILINEAR ,lowercase__ : bool = True ,lowercase__ : bool = True ,**lowercase__ : Optional[Any] ,): super().__init__(**lowercase__ ) __lowercase = size if size is not None else {'''height''': 2_5_6, '''width''': 2_5_6} __lowercase = get_size_dict(lowercase__ ) __lowercase = np.array(lowercase__ ) if clusters is not None else None __lowercase = do_resize __lowercase = size __lowercase = resample __lowercase = do_normalize __lowercase = do_color_quantize def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : np.ndarray ,lowercase__ : Dict[str, int] ,lowercase__ : PILImageResampling = PILImageResampling.BILINEAR ,lowercase__ : Optional[Union[str, ChannelDimension]] = None ,**lowercase__ : str ,): __lowercase = get_size_dict(lowercase__ ) if "height" not in size or "width" not in size: raise ValueError(F"Size dictionary must contain both height and width keys. Got {size.keys()}" ) return resize( lowercase__ ,size=(size['''height'''], size['''width''']) ,resample=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : np.ndarray ,lowercase__ : Optional[Union[str, ChannelDimension]] = None ,): __lowercase = rescale(image=lowercase__ ,scale=1 / 1_2_7.5 ,data_format=lowercase__ ) __lowercase = image - 1 return image def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : ImageInput ,lowercase__ : bool = None ,lowercase__ : Dict[str, int] = None ,lowercase__ : PILImageResampling = None ,lowercase__ : bool = None ,lowercase__ : Optional[bool] = None ,lowercase__ : Optional[Union[List[List[int]], np.ndarray]] = None ,lowercase__ : Optional[Union[str, TensorType]] = None ,lowercase__ : Optional[Union[str, ChannelDimension]] = ChannelDimension.FIRST ,**lowercase__ : Any ,): __lowercase = do_resize if do_resize is not None else self.do_resize __lowercase = size if size is not None else self.size __lowercase = get_size_dict(lowercase__ ) __lowercase = resample if resample is not None else self.resample __lowercase = do_normalize if do_normalize is not None else self.do_normalize __lowercase = do_color_quantize if do_color_quantize is not None else self.do_color_quantize __lowercase = clusters if clusters is not None else self.clusters __lowercase = np.array(lowercase__ ) __lowercase = make_list_of_images(lowercase__ ) if not valid_images(lowercase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_color_quantize and clusters is None: raise ValueError('''Clusters must be specified if do_color_quantize is True.''' ) # All transformations expect numpy arrays. __lowercase = [to_numpy_array(lowercase__ ) for image in images] if do_resize: __lowercase = [self.resize(image=lowercase__ ,size=lowercase__ ,resample=lowercase__ ) for image in images] if do_normalize: __lowercase = [self.normalize(image=lowercase__ ) for image in images] if do_color_quantize: __lowercase = [to_channel_dimension_format(lowercase__ ,ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) __lowercase = np.array(lowercase__ ) __lowercase = color_quantize(lowercase__ ,lowercase__ ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) __lowercase = images.shape[0] __lowercase = images.reshape(lowercase__ ,-1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. __lowercase = list(lowercase__ ) else: __lowercase = [to_channel_dimension_format(lowercase__ ,lowercase__ ) for image in images] __lowercase = {'''input_ids''': images} return BatchFeature(data=lowercase__ ,tensor_type=lowercase__ )
41
from __future__ import annotations import numpy as np def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: return np.maximum(0 , lowerCamelCase_ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
89
0
'''simple docstring''' import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler A_ = 16 A_ = 32 def _UpperCamelCase ( __UpperCamelCase ,__UpperCamelCase = 16 ,__UpperCamelCase = "bert-base-cased" ) -> List[Any]: lowerCamelCase_ = AutoTokenizer.from_pretrained(__UpperCamelCase ) lowerCamelCase_ = load_dataset('glue' ,'mrpc' ) def tokenize_function(__UpperCamelCase ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase_ = tokenizer(examples['sentence1'] ,examples['sentence2'] ,truncation=__UpperCamelCase ,max_length=__UpperCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset lowerCamelCase_ = datasets.map( __UpperCamelCase ,batched=__UpperCamelCase ,remove_columns=['idx', 'sentence1', 'sentence2'] ,load_from_cache_file=__UpperCamelCase ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase_ = tokenized_datasets.rename_column('label' ,'labels' ) def collate_fn(__UpperCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(__UpperCamelCase ,padding='max_length' ,max_length=1_28 ,return_tensors='pt' ) return tokenizer.pad(__UpperCamelCase ,padding='longest' ,return_tensors='pt' ) # Instantiate dataloaders. lowerCamelCase_ = DataLoader( tokenized_datasets['train'] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) lowerCamelCase_ = DataLoader( tokenized_datasets['validation'] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) return train_dataloader, eval_dataloader def _UpperCamelCase ( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Union[str, Any]: model.eval() lowerCamelCase_ = 0 for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase_ = model(**__UpperCamelCase ) lowerCamelCase_ = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times lowerCamelCase_ ,lowerCamelCase_ = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(__UpperCamelCase ) - 1: lowerCamelCase_ = predictions[: len(eval_dataloader.dataset ) - samples_seen] lowerCamelCase_ = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=__UpperCamelCase ,references=__UpperCamelCase ,) lowerCamelCase_ = metric.compute() return eval_metric["accuracy"] def _UpperCamelCase ( __UpperCamelCase ,__UpperCamelCase ) -> List[str]: # Initialize accelerator lowerCamelCase_ = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase_ = config['lr'] lowerCamelCase_ = int(config['num_epochs'] ) lowerCamelCase_ = int(config['seed'] ) lowerCamelCase_ = int(config['batch_size'] ) lowerCamelCase_ = args.model_name_or_path set_seed(__UpperCamelCase ) lowerCamelCase_ ,lowerCamelCase_ = get_dataloaders(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase_ = AutoModelForSequenceClassification.from_pretrained(__UpperCamelCase ,return_dict=__UpperCamelCase ) # Instantiate optimizer lowerCamelCase_ = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) lowerCamelCase_ = optimizer_cls(params=model.parameters() ,lr=__UpperCamelCase ) if accelerator.state.deepspeed_plugin is not None: lowerCamelCase_ = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: lowerCamelCase_ = 1 lowerCamelCase_ = (len(__UpperCamelCase ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): lowerCamelCase_ = get_linear_schedule_with_warmup( optimizer=__UpperCamelCase ,num_warmup_steps=0 ,num_training_steps=__UpperCamelCase ,) else: lowerCamelCase_ = DummyScheduler(__UpperCamelCase ,total_num_steps=__UpperCamelCase ,warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ = accelerator.prepare( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # We need to keep track of how many total steps we have iterated over lowerCamelCase_ = 0 # We also need to keep track of the stating epoch so files are named properly lowerCamelCase_ = 0 lowerCamelCase_ = evaluate.load('glue' ,'mrpc' ) lowerCamelCase_ = num_epochs if args.partial_train_epoch is not None: lowerCamelCase_ = args.partial_train_epoch if args.resume_from_checkpoint: accelerator.load_state(args.resume_from_checkpoint ) lowerCamelCase_ = args.resume_from_checkpoint.split('epoch_' )[1] lowerCamelCase_ = '' for char in epoch_string: if char.isdigit(): state_epoch_num += char else: break lowerCamelCase_ = int(__UpperCamelCase ) + 1 lowerCamelCase_ = evaluation_loop(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) accelerator.print('resumed checkpoint performance:' ,__UpperCamelCase ) accelerator.print('resumed checkpoint\'s scheduler\'s lr:' ,lr_scheduler.get_lr()[0] ) accelerator.print('resumed optimizers\'s lr:' ,optimizer.param_groups[0]['lr'] ) with open(os.path.join(args.output_dir ,f'''state_{starting_epoch-1}.json''' ) ,'r' ) as f: lowerCamelCase_ = json.load(__UpperCamelCase ) assert resumed_state["accuracy"] == accuracy, "Accuracy mismatch, loading from checkpoint failed" assert ( resumed_state["lr"] == lr_scheduler.get_lr()[0] ), "Scheduler learning rate mismatch, loading from checkpoint failed" assert ( resumed_state["optimizer_lr"] == optimizer.param_groups[0]["lr"] ), "Optimizer learning rate mismatch, loading from checkpoint failed" assert resumed_state["epoch"] == starting_epoch - 1, "Epoch mismatch, loading from checkpoint failed" return # Now we train the model lowerCamelCase_ = {} for epoch in range(__UpperCamelCase ,__UpperCamelCase ): model.train() for step, batch in enumerate(__UpperCamelCase ): lowerCamelCase_ = model(**__UpperCamelCase ) lowerCamelCase_ = outputs.loss lowerCamelCase_ = loss / gradient_accumulation_steps accelerator.backward(__UpperCamelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 lowerCamelCase_ = f'''epoch_{epoch}''' lowerCamelCase_ = os.path.join(args.output_dir ,__UpperCamelCase ) accelerator.save_state(__UpperCamelCase ) lowerCamelCase_ = evaluation_loop(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) lowerCamelCase_ = accuracy lowerCamelCase_ = lr_scheduler.get_lr()[0] lowerCamelCase_ = optimizer.param_groups[0]['lr'] lowerCamelCase_ = epoch lowerCamelCase_ = overall_step accelerator.print(f'''epoch {epoch}:''' ,__UpperCamelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir ,f'''state_{epoch}.json''' ) ,'w' ) as f: json.dump(__UpperCamelCase ,__UpperCamelCase ) def _UpperCamelCase ( ) -> str: lowerCamelCase_ = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' ,type=__UpperCamelCase ,default='bert-base-cased' ,help='Path to pretrained model or model identifier from huggingface.co/models.' ,required=__UpperCamelCase ,) parser.add_argument( '--output_dir' ,type=__UpperCamelCase ,default='.' ,help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' ,) parser.add_argument( '--resume_from_checkpoint' ,type=__UpperCamelCase ,default=__UpperCamelCase ,help='If the training should continue from a checkpoint folder.' ,) parser.add_argument( '--partial_train_epoch' ,type=__UpperCamelCase ,default=__UpperCamelCase ,help='If passed, the training will stop after this number of epochs.' ,) parser.add_argument( '--num_epochs' ,type=__UpperCamelCase ,default=2 ,help='Number of train epochs.' ,) lowerCamelCase_ = parser.parse_args() lowerCamelCase_ = {'lr': 2e-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(__UpperCamelCase ,__UpperCamelCase ) if __name__ == "__main__": main()
42
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: # Initialise PyTorch model _lowercase : Optional[int] = TaConfig.from_json_file(lowerCamelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase : Union[str, Any] = TaForConditionalGeneration(lowerCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
89
0
# HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers lowerCAmelCase = float('nan') class _a : def __init__( self: Tuple , UpperCamelCase_: List[Any] ) -> Optional[Any]: """simple docstring""" lowercase__ = sys.stdout lowercase__ = open(UpperCamelCase_ , '''a''' ) def __getattr__( self: List[str] , UpperCamelCase_: str ) -> Optional[Any]: """simple docstring""" return getattr(self.stdout , UpperCamelCase_ ) def lowerCamelCase_ ( self: Dict , UpperCamelCase_: List[str] ) -> str: """simple docstring""" self.stdout.write(UpperCamelCase_ ) # strip tqdm codes self.file.write(re.sub(r'''^.*\r''' , '''''' , UpperCamelCase_ , 0 , re.M ) ) def _a ( SCREAMING_SNAKE_CASE=80 , SCREAMING_SNAKE_CASE=False ): """simple docstring""" lowercase__ = [] # deal with critical env vars lowercase__ = ['''CUDA_VISIBLE_DEVICES'''] for key in env_keys: lowercase__ = os.environ.get(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if val is not None: cmd.append(f'{key}={val}' ) # python executable (not always needed if the script is executable) lowercase__ = sys.executable if full_python_path else sys.executable.split('''/''' )[-1] cmd.append(SCREAMING_SNAKE_CASE ) # now the normal args cmd += list(map(shlex.quote , sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes lowercase__ = [] lowercase__ = '''''' while len(SCREAMING_SNAKE_CASE ) > 0: current_line += f'{cmd.pop(0 )} ' if len(SCREAMING_SNAKE_CASE ) == 0 or len(SCREAMING_SNAKE_CASE ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(SCREAMING_SNAKE_CASE ) lowercase__ = '''''' return "\\\n".join(SCREAMING_SNAKE_CASE ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = re.sub(R'''[\\\n]+''' , ''' ''' , args.base_cmd ) # remove --output_dir if any and set our own lowercase__ = re.sub('''--output_dir\s+[^\s]+''' , '''''' , args.base_cmd ) args.base_cmd += f' --output_dir {output_dir}' # ensure we have --overwrite_output_dir lowercase__ = re.sub('''--overwrite_output_dir\s+''' , '''''' , args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 , 1_00 ) for k in metric_keys} , **{target_metric_key: random.choice([nan, 10.31, 100.2, 55.6_666, 222.22_222_222] )} , ) lowercase__ = subprocess.run(SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE , text=SCREAMING_SNAKE_CASE ) if verbose: print('''STDOUT''' , result.stdout ) print('''STDERR''' , result.stderr ) # save the streams lowercase__ = variation.replace(''' ''' , '''-''' ) with open(Path(SCREAMING_SNAKE_CASE ) / f'log.{prefix}.stdout.txt' , '''w''' ) as f: f.write(result.stdout ) with open(Path(SCREAMING_SNAKE_CASE ) / f'log.{prefix}.stderr.txt' , '''w''' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('''failed''' ) return {target_metric_key: nan} with io.open(f'{output_dir}/all_results.json' , '''r''' , encoding='''utf-8''' ) as f: lowercase__ = json.load(SCREAMING_SNAKE_CASE ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ): """simple docstring""" lowercase__ = [] lowercase__ = [] lowercase__ = f'{id}: {variation:<{longest_variation_len}}' lowercase__ = f'{preamble}: ' lowercase__ = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(SCREAMING_SNAKE_CASE ) , desc=SCREAMING_SNAKE_CASE , leave=SCREAMING_SNAKE_CASE ): lowercase__ = process_run_single( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowercase__ = single_run_metrics[target_metric_key] if not math.isnan(SCREAMING_SNAKE_CASE ): metrics.append(SCREAMING_SNAKE_CASE ) results.append(SCREAMING_SNAKE_CASE ) outcome += "✓" else: outcome += "✘" lowercase__ = f'\33[2K\r{outcome}' if len(SCREAMING_SNAKE_CASE ) > 0: lowercase__ = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} lowercase__ = round(mean_metrics[target_metric_key] , 2 ) lowercase__ = f'{outcome} {mean_target}' if len(SCREAMING_SNAKE_CASE ) > 1: results_str += f' {tuple(round(SCREAMING_SNAKE_CASE , 2 ) for x in results )}' print(SCREAMING_SNAKE_CASE ) lowercase__ = variation return mean_metrics else: print(SCREAMING_SNAKE_CASE ) return {variation_key: variation, target_metric_key: nan} def _a ( ): """simple docstring""" lowercase__ = torch.cuda.get_device_properties(torch.device('''cuda''' ) ) return f'\nDatetime : {datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S" )}\n\nSoftware:\ntransformers: {transformers.__version__}\ntorch : {torch.__version__}\ncuda : {torch.version.cuda}\npython : {platform.python_version()}\n\nHardware:\n{torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB\n' def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = pd.DataFrame(SCREAMING_SNAKE_CASE ) lowercase__ = '''variation''' lowercase__ = '''diff_%''' lowercase__ = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan lowercase__ = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(SCREAMING_SNAKE_CASE ): # as a fallback, use the minimal value as the sentinel lowercase__ = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(SCREAMING_SNAKE_CASE ): lowercase__ = df.apply( lambda SCREAMING_SNAKE_CASE : round(1_00 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 , axis='''columns''' , ) # re-order columns lowercase__ = [variation_key, target_metric_key, diff_key, *report_metric_keys] lowercase__ = df.reindex(SCREAMING_SNAKE_CASE , axis='''columns''' ) # reorder cols # capitalize lowercase__ = df.rename(str.capitalize , axis='''columns''' ) # make the cols as narrow as possible lowercase__ = df.rename(lambda SCREAMING_SNAKE_CASE : c.replace('''_''' , '''<br>''' ) , axis='''columns''' ) lowercase__ = df.rename(lambda SCREAMING_SNAKE_CASE : c.replace('''_''' , '''\n''' ) , axis='''columns''' ) lowercase__ = ['''''', '''Copy between the cut-here-lines and paste as is to github or a forum'''] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=SCREAMING_SNAKE_CASE , floatfmt='''.2f''' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=SCREAMING_SNAKE_CASE , floatfmt='''.2f''' )] print('''\n\n'''.join(SCREAMING_SNAKE_CASE ) ) def _a ( ): """simple docstring""" lowercase__ = argparse.ArgumentParser() parser.add_argument( '''--base-cmd''' , default=SCREAMING_SNAKE_CASE , type=SCREAMING_SNAKE_CASE , required=SCREAMING_SNAKE_CASE , help='''Base cmd''' , ) parser.add_argument( '''--variations''' , default=SCREAMING_SNAKE_CASE , type=SCREAMING_SNAKE_CASE , nargs='''+''' , required=SCREAMING_SNAKE_CASE , help='''Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'''' , ) parser.add_argument( '''--base-variation''' , default=SCREAMING_SNAKE_CASE , type=SCREAMING_SNAKE_CASE , help='''Baseline variation to compare to. if None the minimal target value will be used to compare against''' , ) parser.add_argument( '''--target-metric-key''' , default=SCREAMING_SNAKE_CASE , type=SCREAMING_SNAKE_CASE , required=SCREAMING_SNAKE_CASE , help='''Target metric key in output_dir/all_results.json, e.g., train_samples_per_second''' , ) parser.add_argument( '''--report-metric-keys''' , default='''''' , type=SCREAMING_SNAKE_CASE , help='''Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples''' , ) parser.add_argument( '''--repeat-times''' , default=1 , type=SCREAMING_SNAKE_CASE , help='''How many times to re-run each variation - an average will be reported''' , ) parser.add_argument( '''--output_dir''' , default='''output_benchmark''' , type=SCREAMING_SNAKE_CASE , help='''The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked''' , ) parser.add_argument( '''--verbose''' , default=SCREAMING_SNAKE_CASE , action='''store_true''' , help='''Whether to show the outputs of each run or just the benchmark progress''' , ) lowercase__ = parser.parse_args() lowercase__ = args.output_dir Path(SCREAMING_SNAKE_CASE ).mkdir(exist_ok=SCREAMING_SNAKE_CASE ) lowercase__ = get_base_command(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # split each dimension into its --foo variations lowercase__ = [list(map(str.strip , re.split(R'''\|''' , SCREAMING_SNAKE_CASE ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty lowercase__ = list(map(str.strip , map(''' '''.join , itertools.product(*SCREAMING_SNAKE_CASE ) ) ) ) lowercase__ = max(len(SCREAMING_SNAKE_CASE ) for x in variations ) # split wanted keys lowercase__ = args.report_metric_keys.split() # capture prints into a log file for convenience lowercase__ = f'benchmark-report-{datetime.datetime.now().strftime("%Y-%m-%d-%H-%M-%S" )}.txt' print(f'\nNote: each run\'s output is also logged under {output_dir}/log.*.std*.txt' ) print(f'and this script\'s output is also piped into {report_fn}' ) lowercase__ = Tee(SCREAMING_SNAKE_CASE ) print(f'\n*** Running {len(SCREAMING_SNAKE_CASE )} benchmarks:' ) print(f'Base command: {" ".join(SCREAMING_SNAKE_CASE )}' ) lowercase__ = '''variation''' lowercase__ = [] for id, variation in enumerate(tqdm(SCREAMING_SNAKE_CASE , desc='''Total completion: ''' , leave=SCREAMING_SNAKE_CASE ) ): lowercase__ = base_cmd + variation.split() results.append( process_run( id + 1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , args.target_metric_key , SCREAMING_SNAKE_CASE , args.repeat_times , SCREAMING_SNAKE_CASE , args.verbose , ) ) process_results(SCREAMING_SNAKE_CASE , args.target_metric_key , SCREAMING_SNAKE_CASE , args.base_variation , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
def UpperCamelCase_( lowerCamelCase_ ) -> int: if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: _lowercase : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Tuple = 0 _lowercase : List[str] = 2 while digits < n: index += 1 _lowercase : Optional[int] = len(str(fibonacci(lowerCamelCase_ ) ) ) return index def UpperCamelCase_( lowerCamelCase_ = 1000 ) -> int: return fibonacci_digits_index(lowerCamelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
89
0
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def A_ ( _lowerCAmelCase : dict , _lowerCAmelCase : str , _lowerCAmelCase : set , _lowerCAmelCase : set , _lowerCAmelCase : dict , _lowerCAmelCase : dict , _lowerCAmelCase : PriorityQueue , _lowerCAmelCase : dict , _lowerCAmelCase : float | int , ): """simple docstring""" for nxt, d in graph[v]: if nxt in visited_forward: continue _lowerCamelCase : str = cst_fwd.get(_lowerCAmelCase , np.inf ) _lowerCamelCase : Any = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) _lowerCamelCase : Dict = new_cost_f _lowerCamelCase : str = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: _lowerCamelCase : List[Any] = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def A_ ( _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : dict , _lowerCAmelCase : dict ): """simple docstring""" _lowerCamelCase : Dict = -1 _lowerCamelCase : List[Any] = set() _lowerCamelCase : Optional[int] = set() _lowerCamelCase : List[Any] = {source: 0} _lowerCamelCase : Dict = {destination: 0} _lowerCamelCase : List[str] = {source: None} _lowerCamelCase : Optional[int] = {destination: None} _lowerCamelCase : PriorityQueue[Any] = PriorityQueue() _lowerCamelCase : PriorityQueue[Any] = PriorityQueue() _lowerCamelCase : List[Any] = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): _lowerCamelCase , _lowerCamelCase : List[str] = queue_forward.get() visited_forward.add(_lowerCAmelCase ) _lowerCamelCase , _lowerCamelCase : List[Any] = queue_backward.get() visited_backward.add(_lowerCAmelCase ) _lowerCamelCase : List[Any] = pass_and_relaxation( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) _lowerCamelCase : List[str] = pass_and_relaxation( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: _lowerCamelCase : Any = shortest_distance return shortest_path_distance UpperCAmelCase_ : Optional[Any] = { 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } UpperCAmelCase_ : Dict = { 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
44
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Union[str, Any] = ["ConditionalDetrFeatureExtractor"] SCREAMING_SNAKE_CASE : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Any = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
from math import sqrt def A ( lowercase__ : int ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(sqrt(lowercase__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def A ( lowercase__ : int = 1_0001 ) -> int: UpperCamelCase__ :Dict = 0 UpperCamelCase__ :Union[str, Any] = 1 while count != nth and number < 3: number += 1 if is_prime(lowercase__ ): count += 1 while count != nth: number += 2 if is_prime(lowercase__ ): count += 1 return number if __name__ == "__main__": print(f'''{solution() = }''')
45
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function SCREAMING_SNAKE_CASE : Union[str, Any] = 1.0_5457_1817E-34 # unit of ℏ : J * s SCREAMING_SNAKE_CASE : int = 3E8 # unit of c : m * s^-1 def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> dict[str, float]: if (force, area, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if force < 0: raise ValueError('Magnitude of force can not be negative' ) if distance < 0: raise ValueError('Distance can not be negative' ) if area < 0: raise ValueError('Area can not be negative' ) if force == 0: _lowercase : int = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: _lowercase : List[Any] = (240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: _lowercase : List[Any] = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError('One and only one argument must be 0' ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
89
0
"""simple docstring""" import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def lowerCamelCase_( _lowerCamelCase ) -> Any: '''simple docstring''' _lowerCamelCase : Any = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "decoder.output_projection.weight", "_float_tensor", "encoder.embed_positions._float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(_lowerCamelCase , _lowerCamelCase ) def lowerCamelCase_( _lowerCamelCase ) -> str: '''simple docstring''' _lowerCamelCase, _lowerCamelCase : List[str] = emb.weight.shape _lowerCamelCase : Dict = nn.Linear(_lowerCamelCase , _lowerCamelCase , bias=_lowerCamelCase ) _lowerCamelCase : Any = emb.weight.data return lin_layer def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase=None ) -> Any: '''simple docstring''' _lowerCamelCase : Union[str, Any] = {} for old_key in state_dict.keys(): _lowerCamelCase : List[str] = old_key if "moe_layer.experts." in key: if expert_idx is not None: _lowerCamelCase : Any = key.replace("moe_layer.experts.0" , F"""ffn.experts.expert_{expert_idx}""" ) else: _lowerCamelCase : Optional[Any] = key.replace("moe_layer.experts." , "ffn.experts.expert_" ) if "gate" in key: _lowerCamelCase : int = key.replace(".moe_layer.gate.wg" , ".ffn.router.classifier" ) if "fc2" and "experts" not in key: _lowerCamelCase : Tuple = key.replace(".fc2." , ".ffn.fc2." ) if "fc1" and "experts" not in key: _lowerCamelCase : Union[str, Any] = key.replace(".fc1." , ".ffn.fc1." ) if ".encoder_attn." in key: _lowerCamelCase : Union[str, Any] = key.replace(".encoder_attn." , ".cross_attention." ) if "encoder_attn_layer_norm" in key: _lowerCamelCase : int = key.replace("encoder_attn_layer_norm" , "cross_attention_layer_norm" ) if "final_layer_norm" in key: _lowerCamelCase : Tuple = key.replace("final_layer_norm" , "ff_layer_norm" ) _lowerCamelCase : List[str] = state_dict[old_key] return new_dict def lowerCamelCase_( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = WEIGHTS_NAME ) -> List[str]: '''simple docstring''' _lowerCamelCase : List[str] = [] _lowerCamelCase : List[Any] = 0 os.makedirs(_lowerCamelCase , exist_ok=_lowerCamelCase ) for expert in range(_lowerCamelCase ): _lowerCamelCase : Dict = switch_checkpoint_path + F"""-rank-{expert}.pt""" if os.path.isfile(_lowerCamelCase ): _lowerCamelCase : List[str] = torch.load(_lowerCamelCase )["model"] remove_ignore_keys_(_lowerCamelCase ) _lowerCamelCase : Any = rename_fairseq_keys(_lowerCamelCase , _lowerCamelCase ) _lowerCamelCase : Tuple = os.path.join( _lowerCamelCase , weights_name.replace(".bin" , F"""-{len(_lowerCamelCase )+1:05d}-of-???.bin""" ) ) torch.save(_lowerCamelCase , _lowerCamelCase ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(_lowerCamelCase )[0]].dtype ) # Add the last block _lowerCamelCase : List[str] = os.path.join(_lowerCamelCase , weights_name.replace(".bin" , F"""-{len(_lowerCamelCase )+1:05d}-of-???.bin""" ) ) _lowerCamelCase : List[str] = torch.load(switch_checkpoint_path + "-shared.pt" )["model"] remove_ignore_keys_(_lowerCamelCase ) _lowerCamelCase : int = rename_fairseq_keys(_lowerCamelCase , _lowerCamelCase ) _lowerCamelCase : List[Any] = shared_weights["decoder.embed_tokens.weight"] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(_lowerCamelCase ) == 1: _lowerCamelCase : Union[str, Any] = os.path.join(_lowerCamelCase , _lowerCamelCase ) torch.save(_lowerCamelCase , _lowerCamelCase ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(_lowerCamelCase , _lowerCamelCase ) # Otherwise, let's build the index _lowerCamelCase : Optional[int] = {} for idx, shard in enumerate(_lowerCamelCase ): _lowerCamelCase : int = weights_name.replace(".bin" , F"""-{idx+1:05d}-of-{len(_lowerCamelCase ):05d}.bin""" ) _lowerCamelCase : str = os.path.join(_lowerCamelCase , weights_name.replace(".bin" , F"""-{idx+1:05d}-of-???.bin""" ) ) os.rename(_lowerCamelCase , os.path.join(_lowerCamelCase , _lowerCamelCase ) ) for key in shard: _lowerCamelCase : Optional[Any] = shard_file # Add the metadata _lowerCamelCase : Union[str, Any] = {"total_size": total_size} _lowerCamelCase : Optional[Any] = {"metadata": metadata, "weight_map": weight_map} with open(os.path.join(_lowerCamelCase , _lowerCamelCase ) , "w" , encoding="utf-8" ) as f: _lowerCamelCase : Union[str, Any] = json.dumps(_lowerCamelCase , indent=2 , sort_keys=_lowerCamelCase ) + "\n" f.write(_lowerCamelCase ) return metadata, index if __name__ == "__main__": _lowerCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--nllb_moe_checkpoint_path''', default='''/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000''', type=str, required=False, help='''Path to a directory containing a folder per layer. Follows the original Google format.''', ) parser.add_argument('''--dtype''', default='''float32''', type=str, required=False, help='''dtype of the saved model''') parser.add_argument( '''--pytorch_dump_folder_path''', default='''/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b''', type=str, required=False, help='''Path to the output pytorch model.''', ) _lowerCAmelCase : List[str] = parser.parse_args() _lowerCAmelCase , _lowerCAmelCase : Any = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 128, args.dtype, ) _lowerCAmelCase : List[Any] = NllbMoeConfig.from_pretrained( '''facebook/nllb-200-3.3B''', encoder_sparse_step=4, decoder_sparse_step=4, num_experts=128 ) config.save_pretrained(args.pytorch_dump_folder_path) _lowerCAmelCase : Union[str, Any] = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print('''Done''') model.save_pretrained(args.pytorch_dump_folder_path)
46
def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) _lowercase : List[str] = 0 _lowercase : Optional[int] = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Any = [int(lowerCamelCase_ ) for i in num_string] _lowercase : List[Any] = 1 for i in range(0 , len(lowerCamelCase_ ) ): total *= numbers[i] _lowercase : Optional[Any] = str(lowerCamelCase_ ) steps += 1 return steps def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) _lowercase : Optional[int] = 0 _lowercase : str = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Dict = [int(lowerCamelCase_ ) for i in num_string] _lowercase : Any = 0 for i in range(0 , len(lowerCamelCase_ ) ): total += numbers[i] _lowercase : Dict = str(lowerCamelCase_ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
89
0
import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor SCREAMING_SNAKE_CASE__ = logging.getLogger(__name__) SCREAMING_SNAKE_CASE__ = 50 # max width of layer names SCREAMING_SNAKE_CASE__ = 70 # max width of quantizer names def UpperCAmelCase__ ( lowerCamelCase_ : Dict ): __a : int = parser.add_argument_group('quant_trainer arguments' ) group.add_argument('--wprec' , type=lowerCamelCase_ , default=8 , help='weight precision' ) group.add_argument('--aprec' , type=lowerCamelCase_ , default=8 , help='activation precision' ) group.add_argument('--quant-per-tensor' , action='store_true' , help='per tensor weight scaling' ) group.add_argument('--quant-disable' , action='store_true' , help='disable all quantizers' ) group.add_argument('--quant-disable-embeddings' , action='store_true' , help='disable all embeddings quantizers' ) group.add_argument('--quant-disable-keyword' , type=lowerCamelCase_ , nargs='+' , help='disable quantizers by keyword' ) group.add_argument('--quant-disable-layer-module' , type=lowerCamelCase_ , help='disable quantizers by keyword under layer.' ) group.add_argument('--quant-enable-layer-module' , type=lowerCamelCase_ , help='enable quantizers by keyword under layer' ) group.add_argument('--calibrator' , default='max' , help='which quantization range calibrator to use' ) group.add_argument('--percentile' , default=lowerCamelCase_ , type=lowerCamelCase_ , help='percentile for PercentileCalibrator' ) group.add_argument('--fuse-qkv' , action='store_true' , help='use the same scale factor for qkv' ) group.add_argument('--clip-gelu' , metavar='N' , type=lowerCamelCase_ , help='clip gelu output maximum value to N' ) group.add_argument( '--recalibrate-weights' , action='store_true' , help=( 'recalibrate weight amaxes by taking the max of the weights.' ' amaxes will be computed with the current quantization granularity (axis).' ) , ) def UpperCAmelCase__ ( lowerCamelCase_ : Dict ): if args.calibrator == "max": __a : str = 'max' elif args.calibrator == "percentile": if args.percentile is None: raise ValueError('Specify --percentile when using percentile calibrator' ) __a : int = 'histogram' elif args.calibrator == "mse": __a : List[str] = 'histogram' else: raise ValueError(f'''Invalid calibrator {args.calibrator}''' ) __a : Optional[Any] = QuantDescriptor(num_bits=args.aprec , calib_method=lowerCamelCase_ ) __a : Optional[Any] = QuantDescriptor(num_bits=args.wprec , axis=(None if args.quant_per_tensor else (0,)) ) quant_nn.QuantLinear.set_default_quant_desc_input(lowerCamelCase_ ) quant_nn.QuantLinear.set_default_quant_desc_weight(lowerCamelCase_ ) def UpperCAmelCase__ ( lowerCamelCase_ : Tuple , lowerCamelCase_ : Any , lowerCamelCase_ : Optional[Any]=False , lowerCamelCase_ : Dict=False ): logger.info('Configuring Model for Quantization' ) logger.info(f'''using quantization package {pytorch_quantization.__file__}''' ) if not calib: if args.quant_disable_embeddings: set_quantizer_by_name(lowerCamelCase_ , ['embeddings'] , which='weight' , _disabled=lowerCamelCase_ ) if args.quant_disable: set_quantizer_by_name(lowerCamelCase_ , [''] , _disabled=lowerCamelCase_ ) if args.quant_disable_keyword: set_quantizer_by_name(lowerCamelCase_ , args.quant_disable_keyword , _disabled=lowerCamelCase_ ) if args.quant_disable_layer_module: set_quantizer_by_name(lowerCamelCase_ , [R'layer.\d+.' + args.quant_disable_layer_module] , _disabled=lowerCamelCase_ ) if args.quant_enable_layer_module: set_quantizer_by_name(lowerCamelCase_ , [R'layer.\d+.' + args.quant_enable_layer_module] , _disabled=lowerCamelCase_ ) if args.recalibrate_weights: recalibrate_weights(lowerCamelCase_ ) if args.fuse_qkv: fuse_qkv(lowerCamelCase_ , lowerCamelCase_ ) if args.clip_gelu: clip_gelu(lowerCamelCase_ , args.clip_gelu ) # if args.local_rank in [-1, 0] and not calib: print_quant_summary(lowerCamelCase_ ) def UpperCAmelCase__ ( lowerCamelCase_ : str ): logger.info('Enabling Calibration' ) for name, module in model.named_modules(): if name.endswith('_quantizer' ): if module._calibrator is not None: module.disable_quant() module.enable_calib() else: module.disable() logger.info(f'''{name:80}: {module}''' ) def UpperCAmelCase__ ( lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Dict ): logger.info('Loading calibrated amax' ) for name, module in model.named_modules(): if name.endswith('_quantizer' ): if module._calibrator is not None: if isinstance(module._calibrator , calib.MaxCalibrator ): module.load_calib_amax() else: module.load_calib_amax('percentile' , percentile=args.percentile ) module.enable_quant() module.disable_calib() else: module.enable() model.cuda() print_quant_summary(lowerCamelCase_ ) def UpperCAmelCase__ ( lowerCamelCase_ : int , lowerCamelCase_ : Optional[Any] ): def fusea(lowerCamelCase_ : Tuple , lowerCamelCase_ : Any , lowerCamelCase_ : Optional[int] ): for mod in [qq, qk, qv]: if not hasattr(lowerCamelCase_ , '_amax' ): print(' WARNING: NO AMAX BUFFER' ) return __a : Any = qq._amax.detach().item() __a : Union[str, Any] = qk._amax.detach().item() __a : int = qv._amax.detach().item() __a : List[Any] = max(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) qq._amax.fill_(lowerCamelCase_ ) qk._amax.fill_(lowerCamelCase_ ) qv._amax.fill_(lowerCamelCase_ ) logger.info(f''' q={q:5.2f} k={k:5.2f} v={v:5.2f} -> {amax:5.2f}''' ) for name, mod in model.named_modules(): if name.endswith('.attention.self' ): logger.info(f'''FUSE_QKV: {name:{name_width}}''' ) fusea(mod.matmul_q_input_quantizer , mod.matmul_k_input_quantizer , mod.matmul_v_input_quantizer ) if args.quant_per_tensor: fusea(mod.query._weight_quantizer , mod.key._weight_quantizer , mod.value._weight_quantizer ) def UpperCAmelCase__ ( lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : str ): for name, mod in model.named_modules(): if name.endswith('.output.dense' ) and not name.endswith('attention.output.dense' ): __a : Dict = mod._input_quantizer._amax.data.detach().item() mod._input_quantizer._amax.data.detach().clamp_(max=lowerCamelCase_ ) __a : List[str] = mod._input_quantizer._amax.data.detach().item() logger.info(f'''CLIP_GELU: {name:{name_width}} amax: {amax_init:5.2f} -> {amax:5.2f}''' ) def UpperCAmelCase__ ( lowerCamelCase_ : List[Any] ): for name, mod in model.named_modules(): if hasattr(lowerCamelCase_ , '_weight_quantizer' ) and mod._weight_quantizer.axis is not None: __a : Optional[int] = mod.weight.shape[0] __a : List[str] = mod._weight_quantizer._amax.detach() __a : List[Any] = torch.ones(lowerCamelCase_ , dtype=amax.dtype , device=amax.device ) * amax print(f'''expanding {name} {amax} -> {mod._weight_quantizer._amax}''' ) def UpperCAmelCase__ ( lowerCamelCase_ : Tuple ): for name, mod in model.named_modules(): if hasattr(lowerCamelCase_ , '_weight_quantizer' ): if not hasattr(mod.weight_quantizer , '_amax' ): print('RECALIB: {name:{name_width}} WARNING: NO AMAX BUFFER' ) continue # determine which axes to reduce across # e.g. a 4D tensor quantized per axis 0 should reduce over (1,2,3) __a : Tuple = set() if mod._weight_quantizer.axis is None else set(mod._weight_quantizer.axis ) __a : str = set(range(len(mod.weight.size() ) ) ) - axis_set __a : Union[str, Any] = pytorch_quantization.utils.reduce_amax(mod.weight , axis=lowerCamelCase_ , keepdims=lowerCamelCase_ ).detach() logger.info(f'''RECALIB: {name:{name_width}} {mod._weight_quantizer._amax.flatten()} -> {amax.flatten()}''' ) __a : Union[str, Any] = amax def UpperCAmelCase__ ( lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Optional[Any]=2_5 , lowerCamelCase_ : Dict=1_8_0 , lowerCamelCase_ : int=None ): if ignore is None: __a : Any = [] elif not isinstance(lowerCamelCase_ , lowerCamelCase_ ): __a : int = [ignore] __a : Optional[int] = 0 for name, mod in model.named_modules(): if not hasattr(lowerCamelCase_ , 'weight' ): continue __a : List[str] = max(lowerCamelCase_ , len(lowerCamelCase_ ) ) for name, mod in model.named_modules(): __a : Optional[Any] = getattr(lowerCamelCase_ , '_input_quantizer' , lowerCamelCase_ ) __a : int = getattr(lowerCamelCase_ , '_weight_quantizer' , lowerCamelCase_ ) if not hasattr(lowerCamelCase_ , 'weight' ): continue if type(lowerCamelCase_ ) in ignore: continue if [True for s in ignore if type(lowerCamelCase_ ) is str and s in name]: continue __a : Any = f'''Act:{input_q.extra_repr()}''' __a : str = f'''Wgt:{weight_q.extra_repr()}''' __a : Optional[int] = f'''{name:{name_width}} {act_str} {wgt_str}''' if len(lowerCamelCase_ ) <= line_width: logger.info(lowerCamelCase_ ) else: logger.info(f'''{name:{name_width}} {act_str}''' ) logger.info(f'''{" ":{name_width}} {wgt_str}''' ) def UpperCAmelCase__ ( lowerCamelCase_ : List[str] ): __a : Optional[int] = 0 for name, mod in model.named_modules(): if isinstance(lowerCamelCase_ , pytorch_quantization.nn.TensorQuantizer ): print(f'''{name:80} {mod}''' ) count += 1 print(f'''{count} TensorQuantizers found in model''' ) def UpperCAmelCase__ ( lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : List[Any] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : Optional[int] , lowerCamelCase_ : str ): __a : List[Any] = getattr(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) if quantizer_mod is not None: assert hasattr(lowerCamelCase_ , lowerCamelCase_ ) setattr(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) else: logger.warning(f'''{name} has no {quantizer}''' ) def UpperCAmelCase__ ( lowerCamelCase_ : List[str] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : Any="both" , **lowerCamelCase_ : Any ): __a : Union[str, Any] = f'''Warning: changing {which} quantizers of {name:{qname_width}}''' for k, v in kwargs.items(): s += f''' {k}={v}''' if which in ["input", "both"]: set_quantizer(lowerCamelCase_ , lowerCamelCase_ , '_input_quantizer' , lowerCamelCase_ , lowerCamelCase_ ) if which in ["weight", "both"]: set_quantizer(lowerCamelCase_ , lowerCamelCase_ , '_weight_quantizer' , lowerCamelCase_ , lowerCamelCase_ ) logger.info(lowerCamelCase_ ) def UpperCAmelCase__ ( lowerCamelCase_ : List[Any] , lowerCamelCase_ : Tuple , **lowerCamelCase_ : List[str] ): for name, mod in model.named_modules(): if hasattr(lowerCamelCase_ , '_input_quantizer' ) or hasattr(lowerCamelCase_ , '_weight_quantizer' ): for n in names: if re.search(lowerCamelCase_ , lowerCamelCase_ ): set_quantizers(lowerCamelCase_ , lowerCamelCase_ , **lowerCamelCase_ ) elif name.endswith('_quantizer' ): for n in names: if re.search(lowerCamelCase_ , lowerCamelCase_ ): __a : Dict = f'''Warning: changing {name:{name_width}}''' for k, v in kwargs.items(): s += f''' {k}={v}''' setattr(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) logger.info(lowerCamelCase_ )
47
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: # initialize config if "resnet-50" in model_name: _lowercase : Union[str, Any] = ResNetConfig.from_pretrained('microsoft/resnet-50' ) elif "resnet-101" in model_name: _lowercase : Optional[Any] = ResNetConfig.from_pretrained('microsoft/resnet-101' ) else: raise ValueError('Model name should include either resnet50 or resnet101' ) _lowercase : Tuple = DetrConfig(use_timm_backbone=lowerCamelCase_ , backbone_config=lowerCamelCase_ ) # set label attributes _lowercase : Any = 'panoptic' in model_name if is_panoptic: _lowercase : List[Any] = 250 else: _lowercase : str = 91 _lowercase : List[Any] = 'huggingface/label-files' _lowercase : Any = 'coco-detection-id2label.json' _lowercase : Tuple = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : int = idalabel _lowercase : Any = {v: k for k, v in idalabel.items()} return config, is_panoptic def UpperCamelCase_( lowerCamelCase_ ) -> Any: # here we list all keys to be renamed (original name on the left, our name on the right) _lowercase : List[str] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.conv1.weight', 'backbone.conv_encoder.model.embedder.embedder.convolution.weight') ) rename_keys.append(('backbone.0.body.bn1.weight', 'backbone.conv_encoder.model.embedder.embedder.normalization.weight') ) rename_keys.append(('backbone.0.body.bn1.bias', 'backbone.conv_encoder.model.embedder.embedder.normalization.bias') ) rename_keys.append(('backbone.0.body.bn1.running_mean', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_mean') ) rename_keys.append(('backbone.0.body.bn1.running_var', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_var') ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var''', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var''', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ] ) return rename_keys def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : str = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> str: _lowercase : Any = '' if is_panoptic: _lowercase : Optional[Any] = 'detr.' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowercase : int = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Tuple = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[str] = in_proj_weight[:256, :] _lowercase : Tuple = in_proj_bias[:256] _lowercase : List[Any] = in_proj_weight[256:512, :] _lowercase : Any = in_proj_bias[256:512] _lowercase : int = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _lowercase : str = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Optional[int] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : Union[str, Any] = in_proj_weight[:256, :] _lowercase : Dict = in_proj_bias[:256] _lowercase : Tuple = in_proj_weight[256:512, :] _lowercase : Dict = in_proj_bias[256:512] _lowercase : str = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _lowercase : Tuple = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) _lowercase : Dict = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _lowercase : List[str] = in_proj_weight_cross_attn[:256, :] _lowercase : Tuple = in_proj_bias_cross_attn[:256] _lowercase : str = in_proj_weight_cross_attn[256:512, :] _lowercase : Union[str, Any] = in_proj_bias_cross_attn[256:512] _lowercase : List[Any] = in_proj_weight_cross_attn[-256:, :] _lowercase : Dict = in_proj_bias_cross_attn[-256:] def UpperCamelCase_( ) -> List[Any]: _lowercase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : str = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) return im @torch.no_grad() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=False ) -> List[Any]: _lowercase , _lowercase : int = get_detr_config(lowerCamelCase_ ) # load original model from torch hub _lowercase : int = { 'detr-resnet-50': 'detr_resnet50', 'detr-resnet-101': 'detr_resnet101', } logger.info(F'''Converting model {model_name}...''' ) _lowercase : Optional[Any] = torch.hub.load('facebookresearch/detr' , model_name_to_original_name[model_name] , pretrained=lowerCamelCase_ ).eval() _lowercase : str = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCamelCase_ ): if is_panoptic: _lowercase : str = 'detr.' + src rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCamelCase_ , is_panoptic=lowerCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowercase : List[Any] = 'detr.model.' if is_panoptic else 'model.' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('detr' ) and not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ) ): _lowercase : Tuple = state_dict.pop(lowerCamelCase_ ) _lowercase : int = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val elif key.startswith('bbox_attention' ) or key.startswith('mask_head' ): continue else: _lowercase : Optional[Any] = state_dict.pop(lowerCamelCase_ ) _lowercase : Union[str, Any] = val else: if not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : List[str] = val # finally, create HuggingFace model and load state dict _lowercase : Optional[Any] = DetrForSegmentation(lowerCamelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCamelCase_ ) model.load_state_dict(lowerCamelCase_ ) model.eval() # verify our conversion on an image _lowercase : str = 'coco_panoptic' if is_panoptic else 'coco_detection' _lowercase : Optional[int] = DetrImageProcessor(format=lowerCamelCase_ ) _lowercase : str = processor(images=prepare_img() , return_tensors='pt' ) _lowercase : Tuple = encoding['pixel_values'] _lowercase : int = detr(lowerCamelCase_ ) _lowercase : Tuple = model(lowerCamelCase_ ) assert torch.allclose(outputs.logits , original_outputs['pred_logits'] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs['pred_boxes'] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['pred_masks'] , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info('Uploading PyTorch model and image processor to the hub...' ) model.push_to_hub(F'''nielsr/{model_name}''' ) processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_name", default="detr-resnet-50", type=str, choices=["detr-resnet-50", "detr-resnet-101"], help="Name of the DETR model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the model to the hub or not.") SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class A ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): snake_case__ :str = StableUnCLIPImgaImgPipeline snake_case__ :Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS snake_case__ :Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS snake_case__ :List[str] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess snake_case__ :Optional[int] = frozenset([] ) def __SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" lowerCAmelCase__ = 32 lowerCAmelCase__ = embedder_hidden_size # image encoding components lowerCAmelCase__ = CLIPImageProcessor(crop_size=32 , size=32 ) torch.manual_seed(0 ) lowerCAmelCase__ = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__magic_name__ , projection_dim=__magic_name__ , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) ) # regular denoising components torch.manual_seed(0 ) lowerCAmelCase__ = StableUnCLIPImageNormalizer(embedding_dim=__magic_name__ ) lowerCAmelCase__ = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) lowerCAmelCase__ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) lowerCAmelCase__ = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__magic_name__ , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) lowerCAmelCase__ = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__magic_name__ , layers_per_block=1 , upcast_attention=__magic_name__ , use_linear_projection=__magic_name__ , ) torch.manual_seed(0 ) lowerCAmelCase__ = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.0_0085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=__magic_name__ , steps_offset=1 , ) torch.manual_seed(0 ) lowerCAmelCase__ = AutoencoderKL() lowerCAmelCase__ = { # image encoding components "feature_extractor": feature_extractor, "image_encoder": image_encoder.eval(), # image noising components "image_normalizer": image_normalizer.eval(), "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder.eval(), "unet": unet.eval(), "scheduler": scheduler, "vae": vae.eval(), } return components def __SCREAMING_SNAKE_CASE ( self : int , __magic_name__ : Union[str, Any] , __magic_name__ : Optional[int]=0 , __magic_name__ : int=True ): """simple docstring""" if str(__magic_name__ ).startswith("mps" ): lowerCAmelCase__ = torch.manual_seed(__magic_name__ ) else: lowerCAmelCase__ = torch.Generator(device=__magic_name__ ).manual_seed(__magic_name__ ) lowerCAmelCase__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(__magic_name__ ) ).to(__magic_name__ ) if pil_image: lowerCAmelCase__ = input_image * 0.5 + 0.5 lowerCAmelCase__ = input_image.clamp(0 , 1 ) lowerCAmelCase__ = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() lowerCAmelCase__ = DiffusionPipeline.numpy_to_pil(__magic_name__ )[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def __SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" lowerCAmelCase__ = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase__ = self.get_dummy_components() lowerCAmelCase__ = StableUnCLIPImgaImgPipeline(**__magic_name__ ) lowerCAmelCase__ = sd_pipe.to(__magic_name__ ) sd_pipe.set_progress_bar_config(disable=__magic_name__ ) lowerCAmelCase__ = self.get_dummy_inputs(__magic_name__ ) inputs.update({"image_embeds": None} ) lowerCAmelCase__ = sd_pipe(**__magic_name__ ).images lowerCAmelCase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase__ = np.array([0.3872, 0.7224, 0.5601, 0.4741, 0.6872, 0.5814, 0.4636, 0.3867, 0.5078] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" lowerCAmelCase__ = torch_device in ["cpu", "mps"] self._test_attention_slicing_forward_pass(test_max_difference=__magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" lowerCAmelCase__ = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=__magic_name__ ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__magic_name__ ) @slow @require_torch_gpu class A ( unittest.TestCase ): def __SCREAMING_SNAKE_CASE ( self : List[Any] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" lowerCAmelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) lowerCAmelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy" ) lowerCAmelCase__ = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-l-img2img" , torch_dtype=torch.floataa ) pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCAmelCase__ = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCAmelCase__ = pipe(__magic_name__ , "anime turle" , generator=__magic_name__ , output_type="np" ) lowerCAmelCase__ = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__magic_name__ , __magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" lowerCAmelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) lowerCAmelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy" ) lowerCAmelCase__ = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCAmelCase__ = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCAmelCase__ = pipe(__magic_name__ , "anime turle" , generator=__magic_name__ , output_type="np" ) lowerCAmelCase__ = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__magic_name__ , __magic_name__ ) def __SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" lowerCAmelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png" ) torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCAmelCase__ = StableUnCLIPImgaImgPipeline.from_pretrained( "fusing/stable-unclip-2-1-h-img2img" , torch_dtype=torch.floataa ) lowerCAmelCase__ = pipe.to(__magic_name__ ) pipe.set_progress_bar_config(disable=__magic_name__ ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCAmelCase__ = pipe( __magic_name__ , "anime turtle" , num_inference_steps=2 , output_type="np" , ) lowerCAmelCase__ = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
48
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput SCREAMING_SNAKE_CASE : str = "scheduler_config.json" class _lowerCamelCase( _a ): lowercase_ : Any = 1 lowercase_ : Dict = 2 lowercase_ : Union[str, Any] = 3 lowercase_ : Tuple = 4 lowercase_ : Optional[Any] = 5 @dataclass class _lowerCamelCase( _a ): lowercase_ : jnp.ndarray class _lowerCamelCase: lowercase_ : Union[str, Any] = SCHEDULER_CONFIG_NAME lowercase_ : str = ["""dtype"""] lowercase_ : Dict = [] lowercase_ : int = True @classmethod def UpperCamelCase ( cls, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase=False, **lowerCamelCase, ) -> Optional[int]: """simple docstring""" _lowercase , _lowercase : Optional[int] = cls.load_config( pretrained_model_name_or_path=lowerCamelCase, subfolder=lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase, ) _lowercase , _lowercase : Tuple = cls.from_config(lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase) if hasattr(lowerCamelCase, 'create_state') and getattr(lowerCamelCase, 'has_state', lowerCamelCase): _lowercase : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, **lowerCamelCase) -> Any: """simple docstring""" self.save_config(save_directory=lowerCamelCase, push_to_hub=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return self._get_compatibles() @classmethod def UpperCamelCase ( cls) -> Any: """simple docstring""" _lowercase : Any = list(set([cls.__name__] + cls._compatibles)) _lowercase : Dict = importlib.import_module(__name__.split('.')[0]) _lowercase : Any = [ getattr(lowerCamelCase, lowerCamelCase) for c in compatible_classes_str if hasattr(lowerCamelCase, lowerCamelCase) ] return compatible_classes def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> jnp.ndarray: assert len(lowerCamelCase_ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(lowerCamelCase_ ) - x.ndim) ) , lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=0.9_99 , lowerCamelCase_=jnp.floataa ) -> jnp.ndarray: def alpha_bar(lowerCamelCase_ ): return math.cos((time_step + 0.0_08) / 1.0_08 * math.pi / 2 ) ** 2 _lowercase : List[Any] = [] for i in range(lowerCamelCase_ ): _lowercase : Any = i / num_diffusion_timesteps _lowercase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(lowerCamelCase_ ) / alpha_bar(lowerCamelCase_ ) , lowerCamelCase_ ) ) return jnp.array(lowerCamelCase_ , dtype=lowerCamelCase_ ) @flax.struct.dataclass class _lowerCamelCase: lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray @classmethod def UpperCamelCase ( cls, lowerCamelCase) -> str: """simple docstring""" _lowercase : int = scheduler.config if config.trained_betas is not None: _lowercase : str = jnp.asarray(config.trained_betas, dtype=scheduler.dtype) elif config.beta_schedule == "linear": _lowercase : List[Any] = jnp.linspace(config.beta_start, config.beta_end, config.num_train_timesteps, dtype=scheduler.dtype) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _lowercase : Dict = ( jnp.linspace( config.beta_start**0.5, config.beta_end**0.5, config.num_train_timesteps, dtype=scheduler.dtype) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _lowercase : Optional[int] = betas_for_alpha_bar(config.num_train_timesteps, dtype=scheduler.dtype) else: raise NotImplementedError( F'''beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}''') _lowercase : List[str] = 1.0 - betas _lowercase : Union[str, Any] = jnp.cumprod(lowerCamelCase, axis=0) return cls( alphas=lowerCamelCase, betas=lowerCamelCase, alphas_cumprod=lowerCamelCase, ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : str = state.alphas_cumprod _lowercase : str = alphas_cumprod[timesteps] ** 0.5 _lowercase : Optional[Any] = sqrt_alpha_prod.flatten() _lowercase : Tuple = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) _lowercase : Union[str, Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowercase : Optional[Any] = sqrt_one_minus_alpha_prod.flatten() _lowercase : int = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase , _lowercase : Optional[int] = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: _lowercase , _lowercase : Tuple = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
89
0
"""simple docstring""" from collections import defaultdict from math import gcd def lowercase__ ( snake_case_ :int = 1_500_000 ): __UpperCAmelCase = defaultdict(snake_case_ ) __UpperCAmelCase = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , snake_case_ , 2 ): if gcd(snake_case_ , snake_case_ ) > 1: continue __UpperCAmelCase = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(snake_case_ , limit + 1 , snake_case_ ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(f"""{solution() = }""")
49
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> float: if not nums: raise ValueError('List is empty' ) return sum(lowerCamelCase_ ) / len(lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
89
0
'''simple docstring''' from collections.abc import Generator def A__ ( ): lowerCamelCase__ , lowerCamelCase__ = 0, 1 while True: lowerCamelCase__ , lowerCamelCase__ = b, a + b yield b def A__ ( __lowerCAmelCase : int = 1000 ): lowerCamelCase__ = 1 lowerCamelCase__ = fibonacci_generator() while len(str(next(__lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
50
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def UpperCamelCase_( ) -> List[Any]: _lowercase : int = ArgumentParser('Transformers CLI tool' , usage='transformers-cli <command> [<args>]' ) _lowercase : Optional[Any] = parser.add_subparsers(help='transformers-cli command helpers' ) # Register commands ConvertCommand.register_subcommand(lowerCamelCase_ ) DownloadCommand.register_subcommand(lowerCamelCase_ ) EnvironmentCommand.register_subcommand(lowerCamelCase_ ) RunCommand.register_subcommand(lowerCamelCase_ ) ServeCommand.register_subcommand(lowerCamelCase_ ) UserCommands.register_subcommand(lowerCamelCase_ ) AddNewModelCommand.register_subcommand(lowerCamelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCamelCase_ ) LfsCommands.register_subcommand(lowerCamelCase_ ) PTtoTFCommand.register_subcommand(lowerCamelCase_ ) # Let's go _lowercase : Any = parser.parse_args() if not hasattr(lowerCamelCase_ , 'func' ): parser.print_help() exit(1 ) # Run _lowercase : Optional[int] = args.func(lowerCamelCase_ ) service.run() if __name__ == "__main__": main()
89
0
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' _lowerCamelCase =DiTPipeline _lowerCamelCase =CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS _lowerCamelCase =PipelineTesterMixin.required_optional_params - { "latents", "num_images_per_prompt", "callback", "callback_steps", } _lowerCamelCase =CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS _lowerCamelCase =False def __snake_case ( self : int ): torch.manual_seed(0 ) UpperCAmelCase = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=a__ , activation_fn='''gelu-approximate''' , num_embeds_ada_norm=1000 , norm_type='''ada_norm_zero''' , norm_elementwise_affine=a__ , ) UpperCAmelCase = AutoencoderKL() UpperCAmelCase = DDIMScheduler() UpperCAmelCase = {'''transformer''': transformer.eval(), '''vae''': vae.eval(), '''scheduler''': scheduler} return components def __snake_case ( self : List[Any] , a__ : List[str] , a__ : str=0 ): if str(a__ ).startswith('''mps''' ): UpperCAmelCase = torch.manual_seed(a__ ) else: UpperCAmelCase = torch.Generator(device=a__ ).manual_seed(a__ ) UpperCAmelCase = { '''class_labels''': [1], '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def __snake_case ( self : int ): UpperCAmelCase = '''cpu''' UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**a__ ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) UpperCAmelCase = self.get_dummy_inputs(a__ ) UpperCAmelCase = pipe(**a__ ).images UpperCAmelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) UpperCAmelCase = np.array([0.2_946, 0.6_601, 0.4_329, 0.3_296, 0.4_144, 0.5_319, 0.7_273, 0.5_013, 0.4_457] ) UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(a__ , 1e-3 ) def __snake_case ( self : Optional[int] ): self._test_inference_batch_single_identical(relax_max_difference=a__ , expected_max_diff=1e-3 ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __snake_case ( self : Dict ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @require_torch_gpu @slow class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self : Any ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Any ): UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = DiTPipeline.from_pretrained('''facebook/DiT-XL-2-256''' ) pipe.to('''cuda''' ) UpperCAmelCase = ['''vase''', '''umbrella''', '''white shark''', '''white wolf'''] UpperCAmelCase = pipe.get_label_ids(a__ ) UpperCAmelCase = pipe(a__ , generator=a__ , num_inference_steps=40 , output_type='''np''' ).images for word, image in zip(a__ , a__ ): UpperCAmelCase = load_numpy( f"https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy" ) assert np.abs((expected_image - image).max() ) < 1e-2 def __snake_case ( self : str ): UpperCAmelCase = DiTPipeline.from_pretrained('''facebook/DiT-XL-2-512''' ) UpperCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to('''cuda''' ) UpperCAmelCase = ['''vase''', '''umbrella'''] UpperCAmelCase = pipe.get_label_ids(a__ ) UpperCAmelCase = torch.manual_seed(0 ) UpperCAmelCase = pipe(a__ , generator=a__ , num_inference_steps=25 , output_type='''np''' ).images for word, image in zip(a__ , a__ ): UpperCAmelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' f"/dit/{word}_512.npy" ) assert np.abs((expected_image - image).max() ) < 1e-1
51
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Optional[int] = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def UpperCamelCase ( self, lowerCamelCase=0) -> str: """simple docstring""" _lowercase : Optional[int] = np.random.RandomState(lowerCamelCase) _lowercase : Union[str, Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : List[Any] = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[str] = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Any = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_dummy_inputs() _lowercase : Any = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : Optional[int] = output.images[0, -3:, -3:, -1] _lowercase : int = self.get_dummy_inputs() _lowercase : Union[str, Any] = 3 * [inputs.pop('prompt')] _lowercase : Union[str, Any] = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Tuple = text_inputs['input_ids'] _lowercase : Any = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0] _lowercase : List[Any] = prompt_embeds # forward _lowercase : Union[str, Any] = pipe(**lowerCamelCase) _lowercase : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs() _lowercase : Any = 3 * ['this is a negative prompt'] _lowercase : str = negative_prompt _lowercase : Optional[int] = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : str = output.images[0, -3:, -3:, -1] _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : str = 3 * [inputs.pop('prompt')] _lowercase : Optional[int] = [] for p in [prompt, negative_prompt]: _lowercase : Tuple = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Dict = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0]) _lowercase , _lowercase : str = embeds # forward _lowercase : Dict = pipe(**lowerCamelCase) _lowercase : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = ort.SessionOptions() _lowercase : str = False return options def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'A painting of a squirrel eating a burger' np.random.seed(0) _lowercase : Union[str, Any] = sd_pipe([prompt], guidance_scale=6.0, num_inference_steps=10, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : str = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'open neural network exchange' _lowercase : List[Any] = np.random.RandomState(0) _lowercase : Optional[Any] = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = 'open neural network exchange' _lowercase : str = np.random.RandomState(0) _lowercase : Dict = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[Any] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : List[Any] = 0 def test_callback_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase) -> None: _lowercase : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _lowercase : Any = latents[0, -3:, -3:, -1] _lowercase : Tuple = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _lowercase : List[Any] = latents[0, -3:, -3:, -1] _lowercase : str = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 _lowercase : Any = False _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = 'Andromeda galaxy in a bottle' _lowercase : str = np.random.RandomState(0) pipe( prompt=lowerCamelCase, num_inference_steps=5, guidance_scale=7.5, generator=lowerCamelCase, callback=lowerCamelCase, callback_steps=1, ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) assert isinstance(lowerCamelCase, lowerCamelCase) assert pipe.safety_checker is None _lowercase : Optional[int] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase) _lowercase : Any = OnnxStableDiffusionPipeline.from_pretrained(lowerCamelCase) # sanity check that the pipeline still works assert pipe.safety_checker is None _lowercase : List[str] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None
89
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A = { '''configuration_squeezebert''': [ '''SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SqueezeBertConfig''', '''SqueezeBertOnnxConfig''', ], '''tokenization_squeezebert''': ['''SqueezeBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['''SqueezeBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ '''SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SqueezeBertForMaskedLM''', '''SqueezeBertForMultipleChoice''', '''SqueezeBertForQuestionAnswering''', '''SqueezeBertForSequenceClassification''', '''SqueezeBertForTokenClassification''', '''SqueezeBertModel''', '''SqueezeBertModule''', '''SqueezeBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
52
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : List[Any] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = ["PoolFormerFeatureExtractor"] SCREAMING_SNAKE_CASE : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[Any] = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
89
0
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = ["""image_processor""", """tokenizer"""] a_ = """BlipImageProcessor""" a_ = """AutoTokenizer""" def __init__( self : Optional[int] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : str ) -> Optional[Any]: __lowerCAmelCase = False super().__init__(lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = self.image_processor def __call__( self : List[str] , lowerCAmelCase_ : ImageInput = None , lowerCAmelCase_ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Union[bool, str, PaddingStrategy] = False , lowerCAmelCase_ : Union[bool, str, TruncationStrategy] = None , lowerCAmelCase_ : Optional[int] = None , lowerCAmelCase_ : int = 0 , lowerCAmelCase_ : Optional[int] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Union[str, TensorType]] = None , **lowerCAmelCase_ : Optional[Any] , ) -> BatchEncoding: if images is None and text is None: raise ValueError('You have to specify either images or text.' ) # Get only text if images is None: __lowerCAmelCase = self.tokenizer __lowerCAmelCase = self.tokenizer( text=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , stride=lowerCAmelCase_ , pad_to_multiple_of=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , return_overflowing_tokens=lowerCAmelCase_ , return_special_tokens_mask=lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_length=lowerCAmelCase_ , verbose=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ , ) return text_encoding # add pixel_values __lowerCAmelCase = self.image_processor(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ ) if text is not None: __lowerCAmelCase = self.tokenizer( text=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , max_length=lowerCAmelCase_ , stride=lowerCAmelCase_ , pad_to_multiple_of=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , return_overflowing_tokens=lowerCAmelCase_ , return_special_tokens_mask=lowerCAmelCase_ , return_offsets_mapping=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , return_length=lowerCAmelCase_ , verbose=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ , ) else: __lowerCAmelCase = None if text_encoding is not None: encoding_image_processor.update(lowerCAmelCase_ ) return encoding_image_processor def lowercase ( self : Dict , *lowerCAmelCase_ : str , **lowerCAmelCase_ : Any ) -> Optional[Any]: return self.tokenizer.batch_decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) def lowercase ( self : Optional[Any] , *lowerCAmelCase_ : Optional[int] , **lowerCAmelCase_ : Dict ) -> int: return self.tokenizer.decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def lowercase ( self : int ) -> Optional[int]: __lowerCAmelCase = self.tokenizer.model_input_names __lowerCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
53
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore SCREAMING_SNAKE_CASE : int = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" SCREAMING_SNAKE_CASE : Dict = [file for file in filepaths if file != file.lower()] if upper_files: print(F"{len(upper_files)} files contain uppercase characters:") print("\n".join(upper_files) + "\n") SCREAMING_SNAKE_CASE : List[Any] = [file for file in filepaths if " " in file] if space_files: print(F"{len(space_files)} files contain space characters:") print("\n".join(space_files) + "\n") SCREAMING_SNAKE_CASE : Any = [file for file in filepaths if "-" in file] if hyphen_files: print(F"{len(hyphen_files)} files contain hyphen characters:") print("\n".join(hyphen_files) + "\n") SCREAMING_SNAKE_CASE : str = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"{len(nodir_files)} files are not in a directory:") print("\n".join(nodir_files) + "\n") SCREAMING_SNAKE_CASE : Tuple = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
89
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class A ( __lowercase ): _snake_case =42 _snake_case =42 _snake_case =None class A ( __lowercase , __lowercase ): _snake_case =2 @register_to_config def __init__( self: List[str] , _lowerCAmelCase: float = 0.02 , _lowerCAmelCase: float = 100 , _lowerCAmelCase: float = 1.0_07 , _lowerCAmelCase: float = 80 , _lowerCAmelCase: float = 0.05 , _lowerCAmelCase: float = 50 , ) -> Any: '''simple docstring''' UpperCAmelCase_ =sigma_max # setable values UpperCAmelCase_ =None UpperCAmelCase_ =None UpperCAmelCase_ =None # sigma(t_i) def lowerCAmelCase__ ( self: Any , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: Optional[int] = None ) -> torch.FloatTensor: '''simple docstring''' return sample def lowerCAmelCase__ ( self: List[Any] , _lowerCAmelCase: int , _lowerCAmelCase: Union[str, torch.device] = None ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ =num_inference_steps UpperCAmelCase_ =np.arange(0 , self.num_inference_steps )[::-1].copy() UpperCAmelCase_ =torch.from_numpy(_lowerCAmelCase ).to(_lowerCAmelCase ) UpperCAmelCase_ =[ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] UpperCAmelCase_ =torch.tensor(_lowerCAmelCase , dtype=torch.floataa , device=_lowerCAmelCase ) def lowerCAmelCase__ ( self: List[str] , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: float , _lowerCAmelCase: Optional[torch.Generator] = None ) -> Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: UpperCAmelCase_ =min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: UpperCAmelCase_ =0 # sample eps ~ N(0, S_noise^2 * I) UpperCAmelCase_ =self.config.s_noise * randn_tensor(sample.shape , generator=_lowerCAmelCase ).to(sample.device ) UpperCAmelCase_ =sigma + gamma * sigma UpperCAmelCase_ =sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def lowerCAmelCase__ ( self: Optional[int] , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: float , _lowerCAmelCase: float , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: bool = True , ) -> Union[KarrasVeOutput, Tuple]: '''simple docstring''' UpperCAmelCase_ =sample_hat + sigma_hat * model_output UpperCAmelCase_ =(sample_hat - pred_original_sample) / sigma_hat UpperCAmelCase_ =sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=_lowerCAmelCase , derivative=_lowerCAmelCase , pred_original_sample=_lowerCAmelCase ) def lowerCAmelCase__ ( self: Tuple , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: float , _lowerCAmelCase: float , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: bool = True , ) -> Union[KarrasVeOutput, Tuple]: '''simple docstring''' UpperCAmelCase_ =sample_prev + sigma_prev * model_output UpperCAmelCase_ =(sample_prev - pred_original_sample) / sigma_prev UpperCAmelCase_ =sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=_lowerCAmelCase , derivative=_lowerCAmelCase , pred_original_sample=_lowerCAmelCase ) def lowerCAmelCase__ ( self: int , _lowerCAmelCase: Dict , _lowerCAmelCase: List[str] , _lowerCAmelCase: Any ) -> List[Any]: '''simple docstring''' raise NotImplementedError()
54
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Any: _lowercase : str = 10 _lowercase : List[str] = datasets.Features( { 'tokens': datasets.Sequence(datasets.Value('string' ) ), 'labels': datasets.Sequence(datasets.ClassLabel(names=['negative', 'positive'] ) ), 'answers': datasets.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), 'id': datasets.Value('int64' ), } ) _lowercase : Union[str, Any] = datasets.Dataset.from_dict( { 'tokens': [['foo'] * 5] * n, 'labels': [[1] * 5] * n, 'answers': [{'answer_start': [97], 'text': ['1976']}] * 10, 'id': list(range(lowerCamelCase_ ) ), } , features=lowerCamelCase_ , ) return dataset @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : int = str(tmp_path_factory.mktemp('data' ) / 'file.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return filename # FILE_CONTENT + files SCREAMING_SNAKE_CASE : str = "\\n Text data.\n Second line of data." @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'file.txt' _lowercase : List[str] = FILE_CONTENT with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: import bza _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.bz2' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'file.txt.gz' ) _lowercase : Optional[int] = bytes(lowerCamelCase_ , 'utf-8' ) with gzip.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: if datasets.config.LZ4_AVAILABLE: import lza.frame _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.lz4' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with lza.frame.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: if datasets.config.PY7ZR_AVAILABLE: import pyazr _lowercase : int = tmp_path_factory.mktemp('data' ) / 'file.txt.7z' with pyazr.SevenZipFile(lowerCamelCase_ , 'w' ) as archive: archive.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: import tarfile _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: import lzma _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.xz' _lowercase : int = bytes(lowerCamelCase_ , 'utf-8' ) with lzma.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: import zipfile _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'file.txt.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.zst' _lowercase : Dict = bytes(lowerCamelCase_ , 'utf-8' ) with zstd.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'file.xml' _lowercase : Optional[Any] = textwrap.dedent( '\\n <?xml version="1.0" encoding="UTF-8" ?>\n <tmx version="1.4">\n <header segtype="sentence" srclang="ca" />\n <body>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang="en"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang="en"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang="en"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang="en"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang="en"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>' ) with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] SCREAMING_SNAKE_CASE : Optional[Any] = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } SCREAMING_SNAKE_CASE : Tuple = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] SCREAMING_SNAKE_CASE : Any = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> List[str]: return DATA_DICT_OF_LISTS @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Optional[int] = datasets.Dataset.from_dict(lowerCamelCase_ ) _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.sqlite' ) with contextlib.closing(sqlitea.connect(lowerCamelCase_ ) ) as con: _lowercase : Union[str, Any] = con.cursor() cur.execute('CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)' ) for item in DATA: cur.execute('INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : Tuple = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset2.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : str = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: import bza _lowercase : int = tmp_path_factory.mktemp('data' ) / 'dataset.csv.bz2' with open(lowerCamelCase_ , 'rb' ) as f: _lowercase : int = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(csv_path.replace('.csv' , '.CSV' ) ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(csva_path.replace('.csv' , '.CSV' ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.parquet' ) _lowercase : Optional[Any] = pa.schema( { 'col_1': pa.string(), 'col_2': pa.intaa(), 'col_3': pa.floataa(), } ) with open(lowerCamelCase_ , 'wb' ) as f: _lowercase : List[str] = pq.ParquetWriter(lowerCamelCase_ , schema=lowerCamelCase_ ) _lowercase : Any = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(lowerCamelCase_ ) )] for k in DATA[0]} , schema=lowerCamelCase_ ) writer.write_table(lowerCamelCase_ ) writer.close() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : List[Any] = {'data': DATA} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : Optional[Any] = {'data': DATA_DICT_OF_LISTS} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Dict = str(tmp_path_factory.mktemp('data' ) / 'dataset2.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset_312.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_312: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset-str.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_STR: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: import gzip _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Dict: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Tuple = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Optional[int] = ['0', '1', '2', '3'] _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : str = ['0', '1', '2', '3'] _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : List[Any] = ['0', '1', '2', '3'] _lowercase : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset.abc' with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: _lowercase : List[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.ext.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported.ext' ) ) f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported_2.ext' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = '\n'.join(['First', 'Second\u2029with Unicode new line', 'Third'] ) _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset_with_unicode_new_lines.txt' ) with open(lowerCamelCase_ , 'w' , encoding='utf-8' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Dict: return os.path.join('tests' , 'features' , 'data' , 'test_image_rgb.jpg' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> int: return os.path.join('tests' , 'features' , 'data' , 'test_audio_44100.wav' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'dataset.img.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ).replace('.jpg' , '2.jpg' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data_dir' ) (data_dir / "subdir").mkdir() with open(data_dir / 'subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / 'subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden file with open(data_dir / 'subdir' / '.test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '.subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / '.subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) return data_dir
89
0
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 SCREAMING_SNAKE_CASE :List[Any] = { # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 2048-bit 14: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 3072-bit 15: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 4096-bit 16: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 6144-bit 17: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 8192-bit 18: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, } class UpperCAmelCase : '''simple docstring''' def __init__( self : Optional[int] ,A : int = 14 ): if group not in primes: raise ValueError("Unsupported Group" ) __A = primes[group]["prime"] __A = primes[group]["generator"] __A = int(hexlify(urandom(32 ) ) ,base=16 ) def UpperCamelCase_ ( self : Optional[int] ): return hex(self.__private_key )[2:] def UpperCamelCase_ ( self : List[str] ): __A = pow(self.generator ,self.__private_key ,self.prime ) return hex(A )[2:] def UpperCamelCase_ ( self : List[str] ,A : int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= key <= self.prime - 2 and pow(A ,(self.prime - 1) // 2 ,self.prime ) == 1 ) def UpperCamelCase_ ( self : Optional[int] ,A : str ): __A = int(A ,base=16 ) if not self.is_valid_public_key(A ): raise ValueError("Invalid public key" ) __A = pow(A ,self.__private_key ,self.prime ) return shaaaa(str(A ).encode() ).hexdigest() @staticmethod def UpperCamelCase_ ( A : int ,A : int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= remote_public_key_str <= prime - 2 and pow(A ,(prime - 1) // 2 ,A ) == 1 ) @staticmethod def UpperCamelCase_ ( A : str ,A : str ,A : int = 14 ): __A = int(A ,base=16 ) __A = int(A ,base=16 ) __A = primes[group]["prime"] if not DiffieHellman.is_valid_public_key_static(A ,A ): raise ValueError("Invalid public key" ) __A = pow(A ,A ,A ) return shaaaa(str(A ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
55
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE : str = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : int = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
'''simple docstring''' import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def _a (lowercase__ : Optional[Any] , lowercase__ : int=7 ) -> Tuple: """simple docstring""" __snake_case = None if token is not None: __snake_case = {'Accept': 'application/vnd.github+json', 'Authorization': f'Bearer {token}'} # The id of a workflow (not of a workflow run) __snake_case = '636036' __snake_case = f'https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs' # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += f'?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}' __snake_case = requests.get(lowercase__ , headers=lowercase__ ).json() return result["workflow_runs"] def _a (lowercase__ : List[Any] ) -> List[str]: """simple docstring""" __snake_case = get_daily_ci_runs(lowercase__ ) __snake_case = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": __snake_case = workflow_run['id'] break return workflow_run_id def _a (lowercase__ : Union[str, Any] , lowercase__ : int , lowercase__ : List[Any] ) -> Union[str, Any]: """simple docstring""" __snake_case = get_last_daily_ci_runs(lowercase__ ) if workflow_run_id is not None: __snake_case = get_artifacts_links(worflow_run_id=lowercase__ , token=lowercase__ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: __snake_case = artifacts_links[artifact_name] download_artifact( artifact_name=lowercase__ , artifact_url=lowercase__ , output_dir=lowercase__ , token=lowercase__ ) def _a (lowercase__ : Optional[Any] , lowercase__ : Dict , lowercase__ : int ) -> Dict: """simple docstring""" get_last_daily_ci_artifacts(lowercase__ , lowercase__ , lowercase__ ) __snake_case = {} for artifact_name in artifact_names: __snake_case = os.path.join(lowercase__ , f'{artifact_name}.zip' ) if os.path.isfile(lowercase__ ): __snake_case = {} with zipfile.ZipFile(lowercase__ ) as z: for filename in z.namelist(): if not os.path.isdir(lowercase__ ): # read the file with z.open(lowercase__ ) as f: __snake_case = f.read().decode('UTF-8' ) return results
56
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=7, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=2, lowerCamelCase=99, lowerCamelCase=0, lowerCamelCase=32, lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=0.1, lowerCamelCase=0.1, lowerCamelCase=5_12, lowerCamelCase=2, lowerCamelCase=0.0_2, lowerCamelCase=2, lowerCamelCase=4, lowerCamelCase="last", lowerCamelCase=True, lowerCamelCase=None, lowerCamelCase=0, ) -> str: """simple docstring""" _lowercase : Union[str, Any] = parent _lowercase : Optional[Any] = batch_size _lowercase : List[str] = seq_length _lowercase : int = is_training _lowercase : List[str] = use_input_lengths _lowercase : int = use_token_type_ids _lowercase : Any = use_labels _lowercase : Union[str, Any] = gelu_activation _lowercase : List[str] = sinusoidal_embeddings _lowercase : str = causal _lowercase : Optional[int] = asm _lowercase : Union[str, Any] = n_langs _lowercase : List[Any] = vocab_size _lowercase : Any = n_special _lowercase : Any = hidden_size _lowercase : str = num_hidden_layers _lowercase : Union[str, Any] = num_attention_heads _lowercase : Tuple = hidden_dropout_prob _lowercase : Optional[int] = attention_probs_dropout_prob _lowercase : Union[str, Any] = max_position_embeddings _lowercase : List[str] = type_sequence_label_size _lowercase : Any = initializer_range _lowercase : int = num_labels _lowercase : Optional[int] = num_choices _lowercase : Optional[Any] = summary_type _lowercase : Optional[Any] = use_proj _lowercase : int = scope _lowercase : List[Any] = bos_token_id def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Tuple = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) _lowercase : List[str] = random_attention_mask([self.batch_size, self.seq_length]) _lowercase : int = None if self.use_input_lengths: _lowercase : Dict = ( ids_tensor([self.batch_size], vocab_size=2) + self.seq_length - 2 ) # small variation of seq_length _lowercase : Tuple = None if self.use_token_type_ids: _lowercase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length], self.n_langs) _lowercase : Tuple = None _lowercase : int = None _lowercase : int = None if self.use_labels: _lowercase : str = ids_tensor([self.batch_size], self.type_sequence_label_size) _lowercase : str = ids_tensor([self.batch_size, self.seq_length], self.num_labels) _lowercase : Dict = ids_tensor([self.batch_size], 2).float() _lowercase : Tuple = ids_tensor([self.batch_size], self.num_choices) _lowercase : Dict = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return XLMConfig( vocab_size=self.vocab_size, n_special=self.n_special, emb_dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, gelu_activation=self.gelu_activation, sinusoidal_embeddings=self.sinusoidal_embeddings, asm=self.asm, causal=self.causal, n_langs=self.n_langs, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, summary_type=self.summary_type, use_proj=self.use_proj, num_labels=self.num_labels, bos_token_id=self.bos_token_id, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Tuple: """simple docstring""" _lowercase : List[Any] = XLMModel(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = model(lowerCamelCase, lengths=lowerCamelCase, langs=lowerCamelCase) _lowercase : int = model(lowerCamelCase, langs=lowerCamelCase) _lowercase : Any = model(lowerCamelCase) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[Any]: """simple docstring""" _lowercase : Dict = XLMWithLMHeadModel(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> str: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnsweringSimple(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Dict = model(lowerCamelCase) _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) _lowercase : Any = outputs self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnswering(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = model(lowerCamelCase) _lowercase : List[Any] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, p_mask=lowerCamelCase, ) _lowercase : List[str] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, ) ((_lowercase) , ) : Optional[Any] = result_with_labels.to_tuple() _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) ((_lowercase) , ) : Any = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape, ()) self.parent.assertEqual(result.start_top_log_probs.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual(result.start_top_index.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual( result.end_top_log_probs.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual( result.end_top_index.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual(result.cls_logits.shape, (self.batch_size,)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> int: """simple docstring""" _lowercase : Optional[Any] = XLMForSequenceClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[int] = model(lowerCamelCase) _lowercase : Optional[int] = model(lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[str]: """simple docstring""" _lowercase : Any = self.num_labels _lowercase : str = XLMForTokenClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, attention_mask=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Dict: """simple docstring""" _lowercase : Optional[Any] = self.num_choices _lowercase : Optional[int] = XLMForMultipleChoice(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = input_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : int = token_type_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : Optional[Any] = input_mask.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : List[str] = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices)) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Dict = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[Any] = config_and_inputs _lowercase : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return config, inputs_dict @require_torch class _lowerCamelCase( _a, _a, _a, unittest.TestCase ): lowercase_ : Any = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) lowercase_ : Optional[int] = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable lowercase_ : Union[str, Any] = ( { """feature-extraction""": XLMModel, """fill-mask""": XLMWithLMHeadModel, """question-answering""": XLMForQuestionAnsweringSimple, """text-classification""": XLMForSequenceClassification, """text-generation""": XLMWithLMHeadModel, """token-classification""": XLMForTokenClassification, """zero-shot""": XLMForSequenceClassification, } if is_torch_available() else {} ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast') ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase=False) -> Optional[int]: """simple docstring""" _lowercase : Any = super()._prepare_for_class(lowerCamelCase, lowerCamelCase, return_labels=lowerCamelCase) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": _lowercase : Any = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) _lowercase : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) return inputs_dict def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Union[str, Any] = XLMModelTester(self) _lowercase : List[str] = ConfigTester(self, config_class=lowerCamelCase, emb_dim=37) def UpperCamelCase ( self) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*lowerCamelCase) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*lowerCamelCase) def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> int: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_attentions in attentions], [True] * len(lowerCamelCase)) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_attentions in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : Dict = min_length + idx + 1 _lowercase : int = min_length + idx + 1 _lowercase : Dict = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions], [expected_shape] * len(lowerCamelCase)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> List[Any]: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_hidden_states in hidden_states], [True] * len(lowerCamelCase), ) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_hidden_states in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : int = min_length + idx + 1 _lowercase : int = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states], [expected_shape] * len(lowerCamelCase), ) pass @slow def UpperCamelCase ( self) -> int: """simple docstring""" for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Dict = XLMModel.from_pretrained(lowerCamelCase) self.assertIsNotNone(lowerCamelCase) @require_torch class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMWithLMHeadModel.from_pretrained('xlm-mlm-en-2048') model.to(lowerCamelCase) _lowercase : Union[str, Any] = torch.tensor([[14, 4_47]], dtype=torch.long, device=lowerCamelCase) # the president _lowercase : Any = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference _lowercase : str = model.generate(lowerCamelCase, do_sample=lowerCamelCase) self.assertListEqual(output_ids[0].cpu().numpy().tolist(), lowerCamelCase)
89
0
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() A_ : Any = logging.get_logger(__name__) A_ : List[Any] = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } A_ : Dict = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) -> Tuple: for attribute in key.split('.' ): UpperCamelCase_: Dict = getattr(UpperCAmelCase__ , UpperCAmelCase__ ) if weight_type is not None: UpperCamelCase_: int = getattr(UpperCAmelCase__ , UpperCAmelCase__ ).shape else: UpperCamelCase_: List[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": UpperCamelCase_: Optional[Any] = value elif weight_type == "weight_g": UpperCamelCase_: Any = value elif weight_type == "weight_v": UpperCamelCase_: Optional[int] = value elif weight_type == "bias": UpperCamelCase_: Any = value else: UpperCamelCase_: List[Any] = value logger.info(F'''{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.''' ) def snake_case (UpperCAmelCase__ , UpperCAmelCase__ ) -> str: UpperCamelCase_: Any = [] UpperCamelCase_: str = fairseq_model.state_dict() UpperCamelCase_: int = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight UpperCamelCase_: Optional[int] = None for name, value in fairseq_dict.items(): UpperCamelCase_: Optional[int] = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , hf_model.config.feat_extract_norm == 'group' , ) UpperCamelCase_: Optional[Any] = True elif name.split('.' )[0] == "proj": UpperCamelCase_: Union[str, Any] = fairseq_model.proj UpperCamelCase_: int = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: UpperCamelCase_: int = True if "*" in mapped_key: UpperCamelCase_: List[Any] = name.split(UpperCAmelCase__ )[0].split('.' )[-2] UpperCamelCase_: Tuple = mapped_key.replace('*' , UpperCAmelCase__ ) if "weight_g" in name: UpperCamelCase_: Optional[int] = 'weight_g' elif "weight_v" in name: UpperCamelCase_: List[str] = 'weight_v' elif "bias" in name: UpperCamelCase_: int = 'bias' elif "weight" in name: UpperCamelCase_: Optional[Any] = 'weight' else: UpperCamelCase_: List[str] = None set_recursively(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) continue if not is_used: unused_weights.append(UpperCAmelCase__ ) logger.warning(F'''Unused weights: {unused_weights}''' ) return proj_weight def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) -> Dict: UpperCamelCase_: Union[str, Any] = full_name.split('conv_layers.' )[-1] UpperCamelCase_: Tuple = name.split('.' ) UpperCamelCase_: Dict = int(items[0] ) UpperCamelCase_: str = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) UpperCamelCase_: int = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) UpperCamelCase_: int = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) UpperCamelCase_: Dict = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) UpperCamelCase_: Optional[Any] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(UpperCAmelCase__ ) def snake_case (UpperCAmelCase__ ) -> Optional[Any]: UpperCamelCase_ ,UpperCamelCase_: Optional[Any] = emb.weight.shape UpperCamelCase_: Union[str, Any] = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ , bias=UpperCAmelCase__ ) UpperCamelCase_: Optional[int] = emb.weight.data return lin_layer def snake_case (UpperCAmelCase__ ) -> int: with open(UpperCAmelCase__ , 'r' , encoding='utf-8' ) as f: UpperCamelCase_: str = f.readlines() UpperCamelCase_: List[Any] = [line.split(' ' )[0] for line in lines] UpperCamelCase_: Dict = len(UpperCAmelCase__ ) UpperCamelCase_: Tuple = { '<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3, } vocab_dict.update(dict(zip(UpperCAmelCase__ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , ) -> Tuple: UpperCamelCase_: Dict = WavaVecaConfig.from_pretrained(UpperCAmelCase__ ) UpperCamelCase_: Tuple = SpeechaTextaConfig.from_pretrained( UpperCAmelCase__ , vocab_size=UpperCAmelCase__ , decoder_layers=UpperCAmelCase__ , do_stable_layer_norm=UpperCAmelCase__ ) UpperCamelCase_: Any = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , ) UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_: Optional[Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) UpperCamelCase_: Optional[int] = model[0].eval() # set weights for wav2vec2 encoder UpperCamelCase_: Tuple = WavaVecaModel(UpperCAmelCase__ ) UpperCamelCase_: Any = recursively_load_weights_wavaveca(model.encoder , UpperCAmelCase__ ) UpperCamelCase_: Any = SpeechaTextaForCausalLM(UpperCAmelCase__ ) UpperCamelCase_ ,UpperCamelCase_: Tuple = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=UpperCAmelCase__ ) # set output linear layer unexpected_keys.remove('embed_out' ) UpperCamelCase_: Dict = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(F'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(F'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) UpperCamelCase_: Any = SpeechEncoderDecoderModel(encoder=UpperCAmelCase__ , decoder=UpperCAmelCase__ ) UpperCamelCase_: Any = False # add projection layer UpperCamelCase_: int = nn.Parameter(projection_layer.weight ) UpperCamelCase_: str = nn.Parameter(projection_layer.bias ) UpperCamelCase_: Dict = create_vocab_dict(UpperCAmelCase__ ) with open(os.path.join(UpperCAmelCase__ , 'vocab.json' ) , 'w' ) as fp: json.dump(UpperCAmelCase__ , UpperCAmelCase__ ) UpperCamelCase_: Dict = SpeechaTextaTokenizer(os.path.join(UpperCAmelCase__ , 'vocab.json' ) ) tokenizer.save_pretrained(UpperCAmelCase__ ) UpperCamelCase_: int = hf_wavavec.config.to_dict() UpperCamelCase_: Union[str, Any] = tokenizer.pad_token_id UpperCamelCase_: List[Any] = tokenizer.bos_token_id UpperCamelCase_: Any = tokenizer.eos_token_id UpperCamelCase_: Any = 'speech_to_text_2' UpperCamelCase_: List[Any] = 'wav2vec2' UpperCamelCase_: int = SpeechEncoderDecoderConfig.from_dict(UpperCAmelCase__ ) hf_wavavec.save_pretrained(UpperCAmelCase__ ) feature_extractor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ : Dict = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument( '--encoder_config_path', default='facebook/wav2vec2-large-lv60', type=str, help='Path to hf encoder wav2vec2 checkpoint config', ) parser.add_argument( '--decoder_config_path', default='facebook/s2t-small-mustc-en-fr-st', type=str, help='Path to hf decoder s2t checkpoint config', ) parser.add_argument('--vocab_size', default=10224, type=int, help='Vocab size of decoder') parser.add_argument('--num_decoder_layers', default=7, type=int, help='Number of decoder layers') A_ : Dict = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
57
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.17.0.dev0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/text-classification/requirements.txt") SCREAMING_SNAKE_CASE : Dict = logging.getLogger(__name__) @dataclass class _lowerCamelCase: lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""}, ) lowercase_ : int = field( default=10_24, metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of prediction examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the training data."""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the validation data."""} ) lowercase_ : Optional[str] = field(default=_a, metadata={"""help""": """A csv or a json file containing the test data."""} ) def UpperCamelCase ( self) -> Dict: """simple docstring""" if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError('Need either a GLUE task, a training/validation file or a dataset name.') else: _lowercase : int = self.train_file.split('.')[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." _lowercase : Tuple = self.validation_file.split('.')[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class _lowerCamelCase: lowercase_ : str = field( default=_a, metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""}, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""}, ) lowercase_ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) def UpperCamelCase_( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowercase , _lowercase , _lowercase : Tuple = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowercase , _lowercase , _lowercase : Union[str, Any] = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) _lowercase : Union[str, Any] = training_args.get_process_log_level() logger.setLevel(lowerCamelCase_ ) datasets.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _lowercase : Optional[int] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowercase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. _lowercase : Dict = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. _lowercase : Optional[Any] = {'train': data_args.train_file, 'validation': data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: _lowercase : Tuple = data_args.train_file.split('.' )[-1] _lowercase : int = data_args.test_file.split('.' )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." _lowercase : Any = data_args.test_file else: raise ValueError('Need either a GLUE task or a test file for `do_predict`.' ) for key in data_files.keys(): logger.info(F'''load a local file for {key}: {data_files[key]}''' ) if data_args.train_file.endswith('.csv' ): # Loading a dataset from local csv files _lowercase : str = load_dataset('csv' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files _lowercase : Optional[int] = load_dataset('json' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels _lowercase : Optional[Any] = raw_datasets['train'].features['label'].names _lowercase : Any = len(lowerCamelCase_ ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer _lowercase : str = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=lowerCamelCase_ , ) _lowercase : Tuple = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: _lowercase : int = 'max_length' else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _lowercase : str = False # Some models have set the order of the labels to use, so let's make sure we do use it. _lowercase : List[Any] = {'Refused': 0, 'Entailed': 1} _lowercase : Union[str, Any] = {0: 'Refused', 1: 'Entailed'} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) _lowercase : List[str] = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(lowerCamelCase_ ): # Tokenize the texts def _convert_table_text_to_pandas(lowerCamelCase_ ): _lowercase : int = [_table_row.split('#' ) for _table_row in _table_text.strip('\n' ).split('\n' )] _lowercase : Any = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd _lowercase : List[Any] = examples['statement'] _lowercase : Optional[Any] = list(map(_convert_table_text_to_pandas , examples['table_text'] ) ) _lowercase : Union[str, Any] = tokenizer(lowerCamelCase_ , lowerCamelCase_ , padding=lowerCamelCase_ , max_length=lowerCamelCase_ , truncation=lowerCamelCase_ ) _lowercase : Any = examples['label'] return result with training_args.main_process_first(desc='dataset map pre-processing' ): _lowercase : str = raw_datasets.map( lowerCamelCase_ , batched=lowerCamelCase_ , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on dataset' , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _lowercase : Any = raw_datasets['train'] if data_args.max_train_samples is not None: _lowercase : str = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _lowercase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _lowercase : List[Any] = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError('--do_predict requires a test dataset' ) _lowercase : Optional[int] = raw_datasets['test'] if data_args.max_predict_samples is not None: _lowercase : List[str] = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(lowerCamelCase_ ) ) , 3 ): logger.info(F'''Sample {index} of the training set: {train_dataset[index]}.''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCamelCase_ ): _lowercase : Dict = p.predictions[0] if isinstance(p.predictions , lowerCamelCase_ ) else p.predictions _lowercase : Tuple = np.argmax(lowerCamelCase_ , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _lowercase : Any = default_data_collator elif training_args.fpaa: _lowercase : str = DataCollatorWithPadding(lowerCamelCase_ , pad_to_multiple_of=8 ) else: _lowercase : Optional[Any] = None # Initialize our Trainer _lowercase : List[str] = Trainer( model=lowerCamelCase_ , args=lowerCamelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=lowerCamelCase_ , tokenizer=lowerCamelCase_ , data_collator=lowerCamelCase_ , ) # Training if training_args.do_train: _lowercase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _lowercase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowercase : Optional[Any] = last_checkpoint _lowercase : Optional[Any] = trainer.train(resume_from_checkpoint=lowerCamelCase_ ) _lowercase : List[Any] = train_result.metrics _lowercase : Dict = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCamelCase_ ) ) _lowercase : int = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics('train' , lowerCamelCase_ ) trainer.save_metrics('train' , lowerCamelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _lowercase : Tuple = trainer.evaluate(eval_dataset=lowerCamelCase_ ) _lowercase : Any = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCamelCase_ ) _lowercase : Optional[int] = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.log_metrics('eval' , lowerCamelCase_ ) trainer.save_metrics('eval' , lowerCamelCase_ ) if training_args.do_predict: logger.info('*** Predict ***' ) # Removing the `label` columns because it contains -1 and Trainer won't like that. _lowercase : Any = predict_dataset.remove_columns('label' ) _lowercase : Optional[Any] = trainer.predict(lowerCamelCase_ , metric_key_prefix='predict' ).predictions _lowercase : Union[str, Any] = np.argmax(lowerCamelCase_ , axis=1 ) _lowercase : Dict = os.path.join(training_args.output_dir , 'predict_results_tabfact.txt' ) if trainer.is_world_process_zero(): with open(lowerCamelCase_ , 'w' ) as writer: logger.info('***** Predict Results *****' ) writer.write('index\tprediction\n' ) for index, item in enumerate(lowerCamelCase_ ): _lowercase : List[str] = label_list[item] writer.write(F'''{index}\t{item}\n''' ) _lowercase : str = {'finetuned_from': model_args.model_name_or_path, 'tasks': 'text-classification'} if training_args.push_to_hub: trainer.push_to_hub(**lowerCamelCase_ ) else: trainer.create_model_card(**lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
89
0
"""simple docstring""" import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py __lowerCAmelCase : List[Any] = '''.''' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) __lowerCAmelCase : Optional[int] = [ '''Assert''', '''AssignVariableOp''', '''EmptyTensorList''', '''MergeV2Checkpoints''', '''ReadVariableOp''', '''ResourceGather''', '''RestoreV2''', '''SaveV2''', '''ShardedFilename''', '''StatefulPartitionedCall''', '''StaticRegexFullMatch''', '''VarHandleOp''', ] def __lowerCAmelCase ( __UpperCamelCase : Union[str, Any] , __UpperCamelCase : Any , __UpperCamelCase : Union[str, Any] ): '''simple docstring''' snake_case_ : Any = SavedModel() snake_case_ : str = [] with open(os.path.join(__UpperCamelCase , """utils""" , """tf_ops""" , """onnx.json""" ) ) as f: snake_case_ : str = json.load(__UpperCamelCase )["""opsets"""] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(__UpperCamelCase )] ) with open(__UpperCamelCase , """rb""" ) as f: saved_model.ParseFromString(f.read() ) snake_case_ : Tuple = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want snake_case_ : Optional[Any] = sorted(__UpperCamelCase ) snake_case_ : Any = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(__UpperCamelCase ) if strict and len(__UpperCamelCase ) > 0: raise Exception(F'Found the following incompatible ops for the opset {opset}:\n' + incompatible_ops ) elif len(__UpperCamelCase ) > 0: print(F'Found the following incompatible ops for the opset {opset}:' ) print(*__UpperCamelCase , sep="""\n""" ) else: print(F'The saved model {saved_model_path} can properly be converted with ONNX.' ) if __name__ == "__main__": __lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument('''--saved_model_path''', help='''Path of the saved model to check (the .pb file).''') parser.add_argument( '''--opset''', default=12, type=int, help='''The ONNX opset against which the model has to be tested.''' ) parser.add_argument( '''--framework''', choices=['''onnx'''], default='''onnx''', help='''Frameworks against which to test the saved model.''' ) parser.add_argument( '''--strict''', action='''store_true''', help='''Whether make the checking strict (raise errors) or not (raise warnings)''' ) __lowerCAmelCase : Tuple = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
58
from maths.prime_factors import prime_factors def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): _lowercase : str = F'''Input value of [number={number}] must be an integer''' raise TypeError(lowerCamelCase_ ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(lowerCamelCase_ ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
89
0
import functools import logging import os import sys import threading from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional import huggingface_hub.utils as hf_hub_utils from tqdm import auto as tqdm_lib __A = threading.Lock() __A = None __A = { "debug": logging.DEBUG, "info": logging.INFO, "warning": logging.WARNING, "error": logging.ERROR, "critical": logging.CRITICAL, } __A = logging.WARNING __A = True def lowerCAmelCase_ ( ) -> List[str]: """simple docstring""" lowerCamelCase__: Optional[Any] =os.getenv("TRANSFORMERS_VERBOSITY" , __a ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"""Unknown option TRANSFORMERS_VERBOSITY={env_level_str}, """ F"""has to be one of: { ", ".join(log_levels.keys() ) }""" ) return _default_log_level def lowerCAmelCase_ ( ) -> str: """simple docstring""" return __name__.split("." )[0] def lowerCAmelCase_ ( ) -> logging.Logger: """simple docstring""" return logging.getLogger(_get_library_name() ) def lowerCAmelCase_ ( ) -> None: """simple docstring""" global _default_handler with _lock: if _default_handler: # This library has already configured the library root logger. return lowerCamelCase__: Dict =logging.StreamHandler() # Set sys.stderr as stream. lowerCamelCase__: str =sys.stderr.flush # Apply our default configuration to the library root logger. lowerCamelCase__: List[str] =_get_library_root_logger() library_root_logger.addHandler(_default_handler ) library_root_logger.setLevel(_get_default_logging_level() ) lowerCamelCase__: List[str] =False def lowerCAmelCase_ ( ) -> None: """simple docstring""" global _default_handler with _lock: if not _default_handler: return lowerCamelCase__: Optional[Any] =_get_library_root_logger() library_root_logger.removeHandler(_default_handler ) library_root_logger.setLevel(logging.NOTSET ) lowerCamelCase__: Dict =None def lowerCAmelCase_ ( ) -> List[Any]: """simple docstring""" return log_levels def lowerCAmelCase_ ( __a = None ) -> logging.Logger: """simple docstring""" if name is None: lowerCamelCase__: Any =_get_library_name() _configure_library_root_logger() return logging.getLogger(__a ) def lowerCAmelCase_ ( ) -> int: """simple docstring""" _configure_library_root_logger() return _get_library_root_logger().getEffectiveLevel() def lowerCAmelCase_ ( __a ) -> None: """simple docstring""" _configure_library_root_logger() _get_library_root_logger().setLevel(__a ) def lowerCAmelCase_ ( ) -> Optional[int]: """simple docstring""" return set_verbosity(__a ) def lowerCAmelCase_ ( ) -> Dict: """simple docstring""" return set_verbosity(__a ) def lowerCAmelCase_ ( ) -> Optional[Any]: """simple docstring""" return set_verbosity(__a ) def lowerCAmelCase_ ( ) -> Any: """simple docstring""" return set_verbosity(__a ) def lowerCAmelCase_ ( ) -> None: """simple docstring""" _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().removeHandler(_default_handler ) def lowerCAmelCase_ ( ) -> None: """simple docstring""" _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().addHandler(_default_handler ) def lowerCAmelCase_ ( __a ) -> None: """simple docstring""" _configure_library_root_logger() assert handler is not None _get_library_root_logger().addHandler(__a ) def lowerCAmelCase_ ( __a ) -> None: """simple docstring""" _configure_library_root_logger() assert handler is not None and handler not in _get_library_root_logger().handlers _get_library_root_logger().removeHandler(__a ) def lowerCAmelCase_ ( ) -> None: """simple docstring""" _configure_library_root_logger() lowerCamelCase__: Union[str, Any] =False def lowerCAmelCase_ ( ) -> None: """simple docstring""" _configure_library_root_logger() lowerCamelCase__: Optional[Any] =True def lowerCAmelCase_ ( ) -> None: """simple docstring""" lowerCamelCase__: str =_get_library_root_logger().handlers for handler in handlers: lowerCamelCase__: Union[str, Any] =logging.Formatter("[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s" ) handler.setFormatter(__a ) def lowerCAmelCase_ ( ) -> None: """simple docstring""" lowerCamelCase__: Optional[int] =_get_library_root_logger().handlers for handler in handlers: handler.setFormatter(__a ) def lowerCAmelCase_ ( self , *__a , **__a ) -> int: """simple docstring""" lowerCamelCase__: List[Any] =os.getenv("TRANSFORMERS_NO_ADVISORY_WARNINGS" , __a ) if no_advisory_warnings: return self.warning(*__a , **__a ) __A = warning_advice @functools.lru_cache(__a ) def lowerCAmelCase_ ( self , *__a , **__a ) -> str: """simple docstring""" self.warning(*__a , **__a ) __A = warning_once class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : Dict , *UpperCAmelCase_ : Tuple , **UpperCAmelCase_ : Union[str, Any]) ->Any: # pylint: disable=unused-argument '''simple docstring''' lowerCamelCase__: Tuple =args[0] if args else None def __iter__(self : Any) ->List[str]: '''simple docstring''' return iter(self._iterator) def __getattr__(self : Optional[Any] , UpperCAmelCase_ : List[str]) ->Dict: '''simple docstring''' def empty_fn(*UpperCAmelCase_ : Any , **UpperCAmelCase_ : List[Any]): # pylint: disable=unused-argument return return empty_fn def __enter__(self : Union[str, Any]) ->List[Any]: '''simple docstring''' return self def __exit__(self : List[str] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Any) ->Dict: '''simple docstring''' return class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __call__(self : Dict , *UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : List[Any]) ->List[str]: '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm(*UpperCAmelCase_ , **UpperCAmelCase_) else: return EmptyTqdm(*UpperCAmelCase_ , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , *UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : int) ->Tuple: '''simple docstring''' lowerCamelCase__: Tuple =None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*UpperCAmelCase_ , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Optional[int]) ->Tuple: '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm.get_lock() __A = _tqdm_cls() def lowerCAmelCase_ ( ) -> bool: """simple docstring""" global _tqdm_active return bool(_tqdm_active ) def lowerCAmelCase_ ( ) -> List[Any]: """simple docstring""" global _tqdm_active lowerCamelCase__: str =True hf_hub_utils.enable_progress_bars() def lowerCAmelCase_ ( ) -> Optional[int]: """simple docstring""" global _tqdm_active lowerCamelCase__: Dict =False hf_hub_utils.disable_progress_bars()
59
from __future__ import annotations from typing import Any class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = 0) -> None: """simple docstring""" _lowercase , _lowercase : str = row, column _lowercase : Any = [[default_value for c in range(lowerCamelCase)] for r in range(lowerCamelCase)] def __str__( self) -> str: """simple docstring""" _lowercase : Tuple = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _lowercase : str = 0 for row_vector in self.array: for obj in row_vector: _lowercase : Optional[int] = max(lowerCamelCase, len(str(lowerCamelCase))) _lowercase : List[str] = F'''%{max_element_length}s''' # Make string and return def single_line(lowerCamelCase) -> str: nonlocal string_format_identifier _lowercase : Union[str, Any] = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector) line += "]" return line s += "\n".join(single_line(lowerCamelCase) for row_vector in self.array) return s def __repr__( self) -> str: """simple docstring""" return str(self) def UpperCamelCase ( self, lowerCamelCase) -> bool: """simple docstring""" if not (isinstance(lowerCamelCase, (list, tuple)) and len(lowerCamelCase) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self, lowerCamelCase) -> Any: """simple docstring""" assert self.validate_indicies(lowerCamelCase) return self.array[loc[0]][loc[1]] def __setitem__( self, lowerCamelCase, lowerCamelCase) -> None: """simple docstring""" assert self.validate_indicies(lowerCamelCase) _lowercase : Optional[Any] = value def __add__( self, lowerCamelCase) -> Matrix: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) assert self.row == another.row and self.column == another.column # Add _lowercase : Any = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : int = self[r, c] + another[r, c] return result def __neg__( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : List[str] = -self[r, c] return result def __sub__( self, lowerCamelCase) -> Matrix: """simple docstring""" return self + (-another) def __mul__( self, lowerCamelCase) -> Matrix: """simple docstring""" if isinstance(lowerCamelCase, (int, float)): # Scalar multiplication _lowercase : Dict = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] * another return result elif isinstance(lowerCamelCase, lowerCamelCase): # Matrix multiplication assert self.column == another.row _lowercase : str = Matrix(self.row, another.column) for r in range(self.row): for c in range(another.column): for i in range(self.column): result[r, c] += self[r, i] * another[i, c] return result else: _lowercase : Tuple = F'''Unsupported type given for another ({type(lowerCamelCase)})''' raise TypeError(lowerCamelCase) def UpperCamelCase ( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.column, self.row) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] return result def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> Any: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) and isinstance(lowerCamelCase, lowerCamelCase) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _lowercase : Dict = v.transpose() _lowercase : Any = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def UpperCamelCase_( ) -> None: # a^(-1) _lowercase : Optional[int] = Matrix(3 , 3 , 0 ) for i in range(3 ): _lowercase : int = 1 print(F'''a^(-1) is {ainv}''' ) # u, v _lowercase : Dict = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : Dict = 1, 2, -3 _lowercase : List[Any] = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : int = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCamelCase_ , lowerCamelCase_ )}''' ) def UpperCamelCase_( ) -> None: import doctest doctest.testmod() testa()
89
0
import argparse import math import traceback import dateutil.parser as date_parser import requests def lowerCamelCase_ ( _UpperCamelCase ) -> List[Any]: """simple docstring""" snake_case_ : List[Any] = {} snake_case_ : Optional[int] = job['''started_at'''] snake_case_ : str = job['''completed_at'''] snake_case_ : Dict = date_parser.parse(_UpperCamelCase ) snake_case_ : Tuple = date_parser.parse(_UpperCamelCase ) snake_case_ : str = round((end_datetime - start_datetime).total_seconds() / 60.0 ) snake_case_ : List[str] = start snake_case_ : Optional[int] = end snake_case_ : List[str] = duration_in_min return job_info def lowerCamelCase_ ( _UpperCamelCase , _UpperCamelCase=None ) -> List[Any]: """simple docstring""" snake_case_ : int = None if token is not None: snake_case_ : Optional[int] = {'''Accept''': '''application/vnd.github+json''', '''Authorization''': f'''Bearer {token}'''} snake_case_ : Dict = f'''https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100''' snake_case_ : Optional[Any] = requests.get(_UpperCamelCase , headers=_UpperCamelCase ).json() snake_case_ : List[str] = {} try: job_time.update({job['''name''']: extract_time_from_single_job(_UpperCamelCase ) for job in result['''jobs''']} ) snake_case_ : Optional[int] = math.ceil((result['''total_count'''] - 100) / 100 ) for i in range(_UpperCamelCase ): snake_case_ : Optional[Any] = requests.get(url + f'''&page={i + 2}''' , headers=_UpperCamelCase ).json() job_time.update({job['''name''']: extract_time_from_single_job(_UpperCamelCase ) for job in result['''jobs''']} ) return job_time except Exception: print(f'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') lowerCAmelCase_ = parser.parse_args() lowerCAmelCase_ = get_job_time(args.workflow_run_id) lowerCAmelCase_ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(F'''{k}: {v['duration']}''')
60
import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = int(lowerCamelCase_ ) _lowercase , _lowercase , _lowercase : Optional[Any] = t // 3600, (t // 60) % 60, t % 60 return F'''{h}:{m:02d}:{s:02d}''' if h != 0 else F'''{m:02d}:{s:02d}''' def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=300 ) -> Dict: # docstyle-ignore return F''' <div> {prefix} <progress value=\'{value}\' max=\'{total}\' style=\'width:{width}px; height:20px; vertical-align: middle;\'></progress> {label} </div> ''' def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : int = '<table border="1" class="dataframe">\n' html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += F''' <th>{i}</th>\n''' html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: _lowercase : Any = F'''{elt:.6f}''' if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else str(lowerCamelCase_ ) html_code += F''' <td>{elt}</td>\n''' html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class _lowerCamelCase: lowercase_ : str = 5 lowercase_ : str = 0.2 def __init__( self, lowerCamelCase, lowerCamelCase = None, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = 3_00, ) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = total _lowercase : Optional[int] = '' if prefix is None else prefix _lowercase : Tuple = leave _lowercase : str = parent _lowercase : str = width _lowercase : List[Any] = None _lowercase : List[str] = None _lowercase : Tuple = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, lowerCamelCase = None) -> Dict: """simple docstring""" _lowercase : Any = value if comment is not None: _lowercase : Union[str, Any] = comment if self.last_value is None: _lowercase : Dict = time.time() _lowercase : Tuple = value _lowercase : str = None _lowercase : Optional[int] = self.warmup _lowercase : Optional[Any] = 1 self.update_bar(lowerCamelCase) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for, self.total): if self.first_calls > 0: self.first_calls -= 1 _lowercase : List[str] = time.time() _lowercase : Tuple = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: _lowercase : Dict = self.elapsed_time / (value - self.start_value) else: _lowercase : int = None if value >= self.total: _lowercase : Dict = self.total _lowercase : List[str] = None if not self.leave: self.close() elif self.average_time_per_item is not None: _lowercase : Optional[int] = self.average_time_per_item * (self.total - value) self.update_bar(lowerCamelCase) _lowercase : int = value _lowercase : Tuple = current_time if self.average_time_per_item is None: _lowercase : str = 1 else: _lowercase : int = max(int(self.update_every / self.average_time_per_item), 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None) -> Optional[Any]: """simple docstring""" _lowercase : List[Any] = ' ' * (len(str(self.total)) - len(str(lowerCamelCase))) + str(lowerCamelCase) if self.elapsed_time is None: _lowercase : int = F'''[{spaced_value}/{self.total} : < :''' elif self.predicted_remaining is None: _lowercase : Union[str, Any] = F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)}''' else: _lowercase : Union[str, Any] = ( F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)} <''' F''' {format_time(self.predicted_remaining)}''' ) self.label += F''', {1/self.average_time_per_item:.2f} it/s''' self.label += "]" if self.comment is None or len(self.comment) == 0 else F''', {self.comment}]''' self.display() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: _lowercase : Optional[Any] = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" if self.parent is None and self.output is not None: self.output.update(disp.HTML('')) class _lowerCamelCase( _a ): def __init__( self, lowerCamelCase, lowerCamelCase=None) -> int: """simple docstring""" super().__init__(lowerCamelCase) _lowercase : Optional[Any] = None if column_names is None else [column_names] _lowercase : Any = None def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: _lowercase : Dict = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self, lowerCamelCase) -> Dict: """simple docstring""" if self.inner_table is None: _lowercase : Dict = [list(values.keys()), list(values.values())] else: _lowercase : Tuple = self.inner_table[0] if len(self.inner_table) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(lowerCamelCase) _lowercase : str = columns self.inner_table.append([values[c] for c in columns]) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None, lowerCamelCase=3_00) -> Union[str, Any]: """simple docstring""" _lowercase : List[str] = NotebookProgressBar(lowerCamelCase, prefix=lowerCamelCase, parent=self, width=lowerCamelCase) return self.child_bar def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Optional[Any] = None self.display() class _lowerCamelCase( _a ): def __init__( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = None _lowercase : Dict = None _lowercase : Dict = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Dict: """simple docstring""" _lowercase : Dict = 'Epoch' if args.evaluation_strategy == IntervalStrategy.EPOCH else 'Step' _lowercase : Dict = 0 _lowercase : Tuple = 0 _lowercase : int = [self.first_column] + ['Training Loss'] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append('Validation Loss') _lowercase : Union[str, Any] = NotebookTrainingTracker(state.max_steps, lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" _lowercase : Any = int(state.epoch) if int(state.epoch) == state.epoch else F'''{state.epoch:.2f}''' self.training_tracker.update( state.global_step + 1, comment=F'''Epoch {epoch}/{state.num_train_epochs}''', force_update=self._force_next_update, ) _lowercase : str = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Any: """simple docstring""" if not has_length(lowerCamelCase): return if self.prediction_bar is None: if self.training_tracker is not None: _lowercase : Optional[int] = self.training_tracker.add_child(len(lowerCamelCase)) else: _lowercase : Optional[int] = NotebookProgressBar(len(lowerCamelCase)) self.prediction_bar.update(1) else: self.prediction_bar.update(self.prediction_bar.value + 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Optional[int]: """simple docstring""" if self.prediction_bar is not None: self.prediction_bar.close() _lowercase : Any = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[Any]: """simple docstring""" if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: _lowercase : Dict = {'Training Loss': logs['loss']} # First column is necessarily Step sine we're not in epoch eval strategy _lowercase : List[Any] = state.global_step self.training_tracker.write_line(lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[str]: """simple docstring""" if self.training_tracker is not None: _lowercase : Tuple = {'Training Loss': 'No log', 'Validation Loss': 'No log'} for log in reversed(state.log_history): if "loss" in log: _lowercase : int = log['loss'] break if self.first_column == "Epoch": _lowercase : Union[str, Any] = int(state.epoch) else: _lowercase : Optional[Any] = state.global_step _lowercase : str = 'eval' for k in metrics: if k.endswith('_loss'): _lowercase : str = re.sub(R'\_loss$', '', lowerCamelCase) _lowercase : Tuple = metrics.pop('total_flos', lowerCamelCase) _lowercase : List[str] = metrics.pop('epoch', lowerCamelCase) _lowercase : List[Any] = metrics.pop(F'''{metric_key_prefix}_runtime''', lowerCamelCase) _lowercase : Dict = metrics.pop(F'''{metric_key_prefix}_samples_per_second''', lowerCamelCase) _lowercase : Tuple = metrics.pop(F'''{metric_key_prefix}_steps_per_second''', lowerCamelCase) _lowercase : List[str] = metrics.pop(F'''{metric_key_prefix}_jit_compilation_time''', lowerCamelCase) for k, v in metrics.items(): if k == F'''{metric_key_prefix}_loss''': _lowercase : Union[str, Any] = v else: _lowercase : Optional[Any] = k.split('_') _lowercase : Optional[int] = ' '.join([part.capitalize() for part in splits[1:]]) _lowercase : Tuple = v self.training_tracker.write_line(lowerCamelCase) self.training_tracker.remove_child() _lowercase : str = None # Evaluation takes a long time so we should force the next update. _lowercase : Optional[Any] = True def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" self.training_tracker.update( state.global_step, comment=F'''Epoch {int(state.epoch)}/{state.num_train_epochs}''', force_update=lowerCamelCase) _lowercase : Any = None
89
0
import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'vocab_file': 'spiece.model'} UpperCamelCase = { 'vocab_file': { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model', 'google/bigbird-roberta-large': ( 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model' ), 'google/bigbird-base-trivia-itc': ( 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model' ), } } UpperCamelCase = { 'google/bigbird-roberta-base': 4096, 'google/bigbird-roberta-large': 4096, 'google/bigbird-base-trivia-itc': 4096, } class __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = ["input_ids", "attention_mask"] snake_case__ = [] def __init__( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[str]="<unk>" , SCREAMING_SNAKE_CASE__ : List[str]="<s>" , SCREAMING_SNAKE_CASE__ : Optional[Any]="</s>" , SCREAMING_SNAKE_CASE__ : Tuple="<pad>" , SCREAMING_SNAKE_CASE__ : Any="[SEP]" , SCREAMING_SNAKE_CASE__ : Optional[int]="[MASK]" , SCREAMING_SNAKE_CASE__ : List[Any]="[CLS]" , SCREAMING_SNAKE_CASE__ : Optional[Dict[str, Any]] = None , **SCREAMING_SNAKE_CASE__ : List[Any] , ) -> None: lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else bos_token lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else eos_token lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else unk_token lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else pad_token lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else cls_token lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase__ = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else mask_token lowerCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=SCREAMING_SNAKE_CASE__ , eos_token=SCREAMING_SNAKE_CASE__ , unk_token=SCREAMING_SNAKE_CASE__ , pad_token=SCREAMING_SNAKE_CASE__ , sep_token=SCREAMING_SNAKE_CASE__ , mask_token=SCREAMING_SNAKE_CASE__ , cls_token=SCREAMING_SNAKE_CASE__ , sp_model_kwargs=self.sp_model_kwargs , **SCREAMING_SNAKE_CASE__ , ) lowerCAmelCase__ = vocab_file lowerCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(SCREAMING_SNAKE_CASE__ ) @property def a ( self : List[str] ) -> List[str]: return self.sp_model.get_piece_size() def a ( self : List[str] ) -> Dict: lowerCAmelCase__ = {self.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ) -> Any: lowerCAmelCase__ = self.__dict__.copy() lowerCAmelCase__ = None return state def __setstate__( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Any: lowerCAmelCase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowerCAmelCase__ = {} lowerCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def a ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : str ) -> List[str]: return self.sp_model.encode(SCREAMING_SNAKE_CASE__ , out_type=SCREAMING_SNAKE_CASE__ ) def a ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Tuple: return self.sp_model.piece_to_id(SCREAMING_SNAKE_CASE__ ) def a ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> List[str]: lowerCAmelCase__ = self.sp_model.IdToPiece(SCREAMING_SNAKE_CASE__ ) return token def a ( self : str , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> str: lowerCAmelCase__ = [] lowerCAmelCase__ = "" lowerCAmelCase__ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(SCREAMING_SNAKE_CASE__ ) + token lowerCAmelCase__ = True lowerCAmelCase__ = [] else: current_sub_tokens.append(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = False out_string += self.sp_model.decode(SCREAMING_SNAKE_CASE__ ) return out_string.strip() def a ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = None , SCREAMING_SNAKE_CASE__ : bool = True , **SCREAMING_SNAKE_CASE__ : int , ) -> str: lowerCAmelCase__ = kwargs.pop("use_source_tokenizer" , SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = self.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__ , skip_special_tokens=SCREAMING_SNAKE_CASE__ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCAmelCase__ = [] lowerCAmelCase__ = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(SCREAMING_SNAKE_CASE__ ) ) lowerCAmelCase__ = [] sub_texts.append(SCREAMING_SNAKE_CASE__ ) else: current_sub_text.append(SCREAMING_SNAKE_CASE__ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(SCREAMING_SNAKE_CASE__ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: lowerCAmelCase__ = re.sub(r" (\[(MASK|SEP)\])" , r"\1" , " ".join(SCREAMING_SNAKE_CASE__ ) ) else: lowerCAmelCase__ = "".join(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCAmelCase__ = self.clean_up_tokenization(SCREAMING_SNAKE_CASE__ ) return clean_text else: return text def a ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(SCREAMING_SNAKE_CASE__ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowerCAmelCase__ = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , SCREAMING_SNAKE_CASE__ ) elif not os.path.isfile(self.vocab_file ): with open(SCREAMING_SNAKE_CASE__ , "wb" ) as fi: lowerCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(SCREAMING_SNAKE_CASE__ ) return (out_vocab_file,) def a ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase__ = [self.cls_token_id] lowerCAmelCase__ = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def a ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE__ , token_ids_a=SCREAMING_SNAKE_CASE__ , already_has_special_tokens=SCREAMING_SNAKE_CASE__ ) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] def a ( self : Optional[int] , SCREAMING_SNAKE_CASE__ : List[int] , SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None ) -> List[int]: lowerCAmelCase__ = [self.sep_token_id] lowerCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
61
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 18, 2] _lowercase : Tuple = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Any = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Dict = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: _lowercase : Any = [3, 3, 3, 3] _lowercase : Any = [5, 5, 5, 5] elif "fl4" in model_name: _lowercase : Dict = [4, 4, 4, 4] _lowercase : Tuple = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: _lowercase : str = [3, 3, 3, 3] if "lrf" in model_name: _lowercase : Optional[int] = [3, 3, 3, 3] else: _lowercase : Dict = [2, 2, 2, 2] if "tiny" in model_name: _lowercase : List[str] = 96 elif "small" in model_name: _lowercase : Dict = 96 elif "base" in model_name: _lowercase : Optional[int] = 128 elif "large" in model_name: _lowercase : List[Any] = 192 elif "xlarge" in model_name: _lowercase : Optional[Any] = 256 elif "huge" in model_name: _lowercase : Dict = 352 # set label information _lowercase : int = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: _lowercase : str = 'imagenet-22k-id2label.json' else: _lowercase : Tuple = 'imagenet-1k-id2label.json' _lowercase : Union[str, Any] = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : Any = {v: k for k, v in idalabel.items()} _lowercase : Optional[Any] = FocalNetConfig( embed_dim=lowerCamelCase_ , depths=lowerCamelCase_ , focal_levels=lowerCamelCase_ , focal_windows=lowerCamelCase_ , use_conv_embed=lowerCamelCase_ , idalabel=lowerCamelCase_ , labelaid=lowerCamelCase_ , use_post_layernorm=lowerCamelCase_ , use_layerscale=lowerCamelCase_ , ) return config def UpperCamelCase_( lowerCamelCase_ ) -> Any: if "patch_embed.proj" in name: _lowercase : Optional[Any] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: _lowercase : str = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: _lowercase : Any = 'encoder.' + name if "encoder.layers" in name: _lowercase : int = name.replace('encoder.layers' , 'encoder.stages' ) if "downsample.proj" in name: _lowercase : Tuple = name.replace('downsample.proj' , 'downsample.projection' ) if "blocks" in name: _lowercase : str = name.replace('blocks' , 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: _lowercase : List[str] = name.replace('modulation.f' , 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: _lowercase : int = name.replace('modulation.h' , 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: _lowercase : Any = name.replace('modulation.proj' , 'modulation.projection_out' ) if name == "norm.weight": _lowercase : Any = 'layernorm.weight' if name == "norm.bias": _lowercase : Tuple = 'layernorm.bias' if "head" in name: _lowercase : Optional[int] = name.replace('head' , 'classifier' ) else: _lowercase : Optional[int] = 'focalnet.' + name return name def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False ) -> str: # fmt: off _lowercase : Dict = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on _lowercase : Dict = model_name_to_url[model_name] print('Checkpoint URL: ' , lowerCamelCase_ ) _lowercase : List[str] = torch.hub.load_state_dict_from_url(lowerCamelCase_ , map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[int] = val _lowercase : Union[str, Any] = get_focalnet_config(lowerCamelCase_ ) _lowercase : Optional[Any] = FocalNetForImageClassification(lowerCamelCase_ ) model.eval() # load state dict model.load_state_dict(lowerCamelCase_ ) # verify conversion _lowercase : Optional[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : Any = BitImageProcessor( do_resize=lowerCamelCase_ , size={'shortest_edge': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=lowerCamelCase_ , crop_size=224 , do_normalize=lowerCamelCase_ , image_mean=lowerCamelCase_ , image_std=lowerCamelCase_ , ) _lowercase : List[str] = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) _lowercase : List[Any] = processor(images=lowerCamelCase_ , return_tensors='pt' ) _lowercase : str = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ), ] ) _lowercase : List[str] = image_transforms(lowerCamelCase_ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , lowerCamelCase_ , atol=1e-4 ) _lowercase : Dict = model(**lowerCamelCase_ ) _lowercase : int = outputs.logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) print('First values of logits:' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": _lowercase : Optional[Any] = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": _lowercase : int = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": _lowercase : str = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": _lowercase : Any = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": _lowercase : List[Any] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": _lowercase : int = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3] , lowerCamelCase_ , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
def lowerCamelCase__ ( lowercase , lowercase ): """simple docstring""" return 1 if input_a == input_a else 0 def lowerCamelCase__ ( ): """simple docstring""" assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
62
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Any = { "ut/deta": "https://huggingface.co/ut/deta/resolve/main/config.json", } class _lowerCamelCase( _a ): lowercase_ : Any = """deta""" lowercase_ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self, lowerCamelCase=None, lowerCamelCase=9_00, lowerCamelCase=20_48, lowerCamelCase=6, lowerCamelCase=20_48, lowerCamelCase=8, lowerCamelCase=6, lowerCamelCase=10_24, lowerCamelCase=8, lowerCamelCase=0.0, lowerCamelCase=True, lowerCamelCase="relu", lowerCamelCase=2_56, lowerCamelCase=0.1, lowerCamelCase=0.0, lowerCamelCase=0.0, lowerCamelCase=0.0_2, lowerCamelCase=1.0, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase="sine", lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=4, lowerCamelCase=True, lowerCamelCase=3_00, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=1, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=0.1, lowerCamelCase=0.2_5, **lowerCamelCase, ) -> Any: """simple docstring""" if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.') _lowercase : List[Any] = CONFIG_MAPPING['resnet'](out_features=['stage2', 'stage3', 'stage4']) else: if isinstance(lowerCamelCase, lowerCamelCase): _lowercase : Dict = backbone_config.pop('model_type') _lowercase : int = CONFIG_MAPPING[backbone_model_type] _lowercase : Union[str, Any] = config_class.from_dict(lowerCamelCase) _lowercase : Union[str, Any] = backbone_config _lowercase : Any = num_queries _lowercase : Union[str, Any] = max_position_embeddings _lowercase : Union[str, Any] = d_model _lowercase : Optional[int] = encoder_ffn_dim _lowercase : Optional[int] = encoder_layers _lowercase : Optional[Any] = encoder_attention_heads _lowercase : Optional[Any] = decoder_ffn_dim _lowercase : Dict = decoder_layers _lowercase : Tuple = decoder_attention_heads _lowercase : Union[str, Any] = dropout _lowercase : Optional[Any] = attention_dropout _lowercase : int = activation_dropout _lowercase : Tuple = activation_function _lowercase : List[Any] = init_std _lowercase : Union[str, Any] = init_xavier_std _lowercase : int = encoder_layerdrop _lowercase : Optional[int] = auxiliary_loss _lowercase : Dict = position_embedding_type # deformable attributes _lowercase : Any = num_feature_levels _lowercase : str = encoder_n_points _lowercase : Any = decoder_n_points _lowercase : List[str] = two_stage _lowercase : Dict = two_stage_num_proposals _lowercase : Any = with_box_refine _lowercase : List[Any] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.') # Hungarian matcher _lowercase : List[Any] = class_cost _lowercase : Optional[int] = bbox_cost _lowercase : str = giou_cost # Loss coefficients _lowercase : Optional[int] = mask_loss_coefficient _lowercase : int = dice_loss_coefficient _lowercase : List[Any] = bbox_loss_coefficient _lowercase : Optional[Any] = giou_loss_coefficient _lowercase : str = eos_coefficient _lowercase : int = focal_alpha super().__init__(is_encoder_decoder=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.d_model def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : int = copy.deepcopy(self.__dict__) _lowercase : Optional[int] = self.backbone_config.to_dict() _lowercase : Optional[Any] = self.__class__.model_type return output
89
0
from math import asin, atan, cos, radians, sin, sqrt, tan a : Tuple = 6_37_81_37.0 a : Any = 6_35_67_52.31_42_45 a : Tuple = 6_378_137 def lowerCamelCase__ ( __lowerCamelCase : float , __lowerCamelCase : float , __lowerCamelCase : float , __lowerCamelCase : float ): __UpperCAmelCase : Union[str, Any] = (AXIS_A - AXIS_B) / AXIS_A __UpperCAmelCase : List[str] = atan((1 - flattening) * tan(radians(__lowerCamelCase ) ) ) __UpperCAmelCase : Optional[Any] = atan((1 - flattening) * tan(radians(__lowerCamelCase ) ) ) __UpperCAmelCase : Optional[Any] = radians(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = radians(__lowerCamelCase ) # Equation __UpperCAmelCase : Any = sin((phi_a - phi_a) / 2 ) __UpperCAmelCase : Any = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda __UpperCAmelCase : List[Any] = sqrt(sin_sq_phi + (cos(__lowerCamelCase ) * cos(__lowerCamelCase ) * sin_sq_lambda) ) return 2 * RADIUS * asin(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
63
from __future__ import annotations import numpy as np def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: return np.maximum(0 , lowerCamelCase_ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
89
0
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class _lowerCamelCase : @property def UpperCamelCase_ ( self ) -> Tuple: return self.get_dummy_input() @property def UpperCamelCase_ ( self ) -> Optional[int]: if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f'\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.' ) def UpperCamelCase_ ( self , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase=False , lowerCAmelCase=False , ) -> str: SCREAMING_SNAKE_CASE__: str= 4 SCREAMING_SNAKE_CASE__: List[str]= 32 SCREAMING_SNAKE_CASE__: Any= (32, 32) SCREAMING_SNAKE_CASE__: int= torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__: List[Any]= torch.device(lowerCAmelCase ) SCREAMING_SNAKE_CASE__: List[str]= (batch_size, num_channels) + sizes SCREAMING_SNAKE_CASE__: Union[str, Any]= randn_tensor(lowerCAmelCase , generator=lowerCAmelCase , device=lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Optional[Any]= {'''hidden_states''': hidden_states} if include_temb: SCREAMING_SNAKE_CASE__: Optional[Any]= 128 SCREAMING_SNAKE_CASE__: Union[str, Any]= randn_tensor((batch_size, temb_channels) , generator=lowerCAmelCase , device=lowerCAmelCase ) if include_res_hidden_states_tuple: SCREAMING_SNAKE_CASE__: List[Any]= torch.manual_seed(1 ) SCREAMING_SNAKE_CASE__: Optional[Any]= (randn_tensor(lowerCAmelCase , generator=lowerCAmelCase , device=lowerCAmelCase ),) if include_encoder_hidden_states: SCREAMING_SNAKE_CASE__: Optional[Any]= floats_tensor((batch_size, 32, 32) ).to(lowerCAmelCase ) if include_skip_sample: SCREAMING_SNAKE_CASE__: int= randn_tensor(((batch_size, 3) + sizes) , generator=lowerCAmelCase , device=lowerCAmelCase ) return dummy_input def UpperCamelCase_ ( self ) -> Optional[Any]: SCREAMING_SNAKE_CASE__: int= { '''in_channels''': 32, '''out_channels''': 32, '''temb_channels''': 128, } if self.block_type == "up": SCREAMING_SNAKE_CASE__: Optional[Any]= 32 if self.block_type == "mid": init_dict.pop('''out_channels''' ) SCREAMING_SNAKE_CASE__: List[str]= self.dummy_input return init_dict, inputs_dict def UpperCamelCase_ ( self , lowerCAmelCase ) -> List[str]: SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__: List[Any]= self.prepare_init_args_and_inputs_for_common() SCREAMING_SNAKE_CASE__: Any= self.block_class(**lowerCAmelCase ) unet_block.to(lowerCAmelCase ) unet_block.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__: Dict= unet_block(**lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ): SCREAMING_SNAKE_CASE__: Tuple= output[0] self.assertEqual(output.shape , self.output_shape ) SCREAMING_SNAKE_CASE__: List[str]= output[0, -1, -3:, -3:] SCREAMING_SNAKE_CASE__: Tuple= torch.tensor(lowerCAmelCase ).to(lowerCAmelCase ) assert torch_all_close(output_slice.flatten() , lowerCAmelCase , atol=5e-3 ) @unittest.skipIf(torch_device == '''mps''' , '''Training is not supported in mps''' ) def UpperCamelCase_ ( self ) -> Optional[int]: SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__: str= self.prepare_init_args_and_inputs_for_common() SCREAMING_SNAKE_CASE__: Tuple= self.block_class(**lowerCAmelCase ) model.to(lowerCAmelCase ) model.train() SCREAMING_SNAKE_CASE__: Dict= model(**lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ): SCREAMING_SNAKE_CASE__: Any= output[0] SCREAMING_SNAKE_CASE__: List[Any]= torch.device(lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Any= randn_tensor(output.shape , device=lowerCAmelCase ) SCREAMING_SNAKE_CASE__: Any= torch.nn.functional.mse_loss(lowerCAmelCase , lowerCAmelCase ) loss.backward()
64
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: # Initialise PyTorch model _lowercase : Optional[int] = TaConfig.from_json_file(lowerCamelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase : Union[str, Any] = TaForConditionalGeneration(lowerCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
89
0
"""simple docstring""" from collections.abc import Sequence def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' return sum(c * (x**i) for i, c in enumerate(__UpperCamelCase ) ) def lowerCAmelCase ( __UpperCamelCase , __UpperCamelCase ): '''simple docstring''' UpperCAmelCase__ : List[str] = 0.0 for coeff in reversed(__UpperCamelCase ): UpperCAmelCase__ : Optional[int] = result * x + coeff return result if __name__ == "__main__": __UpperCAmelCase = (0.0, 0.0, 5.0, 9.3, 7.0) __UpperCAmelCase = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
65
def UpperCamelCase_( lowerCamelCase_ ) -> int: if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: _lowercase : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Tuple = 0 _lowercase : List[str] = 2 while digits < n: index += 1 _lowercase : Optional[int] = len(str(fibonacci(lowerCamelCase_ ) ) ) return index def UpperCamelCase_( lowerCamelCase_ = 1000 ) -> int: return fibonacci_digits_index(lowerCamelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
89
0
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Union[str, Any] = ["ConditionalDetrFeatureExtractor"] SCREAMING_SNAKE_CASE : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Any = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
import os def SCREAMING_SNAKE_CASE__ ( ) -> List[Any]: with open(os.path.dirname(snake_case__ ) + '/grid.txt' ) as f: _lowercase = [] # noqa: E741 for _ in range(20 ): l.append([int(snake_case__ ) for x in f.readline().split()] ) _lowercase = 0 # right for i in range(20 ): for j in range(17 ): _lowercase = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: _lowercase = temp # down for i in range(17 ): for j in range(20 ): _lowercase = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: _lowercase = temp # diagonal 1 for i in range(17 ): for j in range(17 ): _lowercase = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: _lowercase = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): _lowercase = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: _lowercase = temp return maximum if __name__ == "__main__": print(solution())
67
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function SCREAMING_SNAKE_CASE : Union[str, Any] = 1.0_5457_1817E-34 # unit of ℏ : J * s SCREAMING_SNAKE_CASE : int = 3E8 # unit of c : m * s^-1 def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> dict[str, float]: if (force, area, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if force < 0: raise ValueError('Magnitude of force can not be negative' ) if distance < 0: raise ValueError('Distance can not be negative' ) if area < 0: raise ValueError('Area can not be negative' ) if force == 0: _lowercase : int = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: _lowercase : List[Any] = (240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: _lowercase : List[Any] = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError('One and only one argument must be 0' ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
89
0
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=5_12, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def lowercase__ ( A_: List[Any] ) -> Union[str, Any]: """simple docstring""" if string == "True": return True elif string == "False": return False else: raise ValueError(F'''could not parse string as bool {string}''' ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) __A = parser.parse_args() __A = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
68
def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) _lowercase : List[str] = 0 _lowercase : Optional[int] = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Any = [int(lowerCamelCase_ ) for i in num_string] _lowercase : List[Any] = 1 for i in range(0 , len(lowerCamelCase_ ) ): total *= numbers[i] _lowercase : Optional[Any] = str(lowerCamelCase_ ) steps += 1 return steps def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) _lowercase : Optional[int] = 0 _lowercase : str = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Dict = [int(lowerCamelCase_ ) for i in num_string] _lowercase : Any = 0 for i in range(0 , len(lowerCamelCase_ ) ): total += numbers[i] _lowercase : Dict = str(lowerCamelCase_ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
89
0
'''simple docstring''' import math def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: __snake_case = [True] * n __snake_case = False __snake_case = False __snake_case = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): __snake_case = i * 2 while index < n: __snake_case = False __snake_case = index + i __snake_case = [2] for i in range(3 , _UpperCAmelCase , 2 ): if is_prime[i]: primes.append(_UpperCAmelCase ) return primes def __UpperCAmelCase ( _UpperCAmelCase : int = 99_99_66_66_33_33 ) -> int: __snake_case = math.floor(math.sqrt(_UpperCAmelCase ) ) + 1_00 __snake_case = prime_sieve(_UpperCAmelCase ) __snake_case = 0 __snake_case = 0 __snake_case = primes[prime_index] while (last_prime**2) <= limit: __snake_case = primes[prime_index + 1] __snake_case = last_prime**2 __snake_case = next_prime**2 # Get numbers divisible by lps(current) __snake_case = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) __snake_case = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps __snake_case = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair __snake_case = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
69
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: # initialize config if "resnet-50" in model_name: _lowercase : Union[str, Any] = ResNetConfig.from_pretrained('microsoft/resnet-50' ) elif "resnet-101" in model_name: _lowercase : Optional[Any] = ResNetConfig.from_pretrained('microsoft/resnet-101' ) else: raise ValueError('Model name should include either resnet50 or resnet101' ) _lowercase : Tuple = DetrConfig(use_timm_backbone=lowerCamelCase_ , backbone_config=lowerCamelCase_ ) # set label attributes _lowercase : Any = 'panoptic' in model_name if is_panoptic: _lowercase : List[Any] = 250 else: _lowercase : str = 91 _lowercase : List[Any] = 'huggingface/label-files' _lowercase : Any = 'coco-detection-id2label.json' _lowercase : Tuple = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : int = idalabel _lowercase : Any = {v: k for k, v in idalabel.items()} return config, is_panoptic def UpperCamelCase_( lowerCamelCase_ ) -> Any: # here we list all keys to be renamed (original name on the left, our name on the right) _lowercase : List[str] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.conv1.weight', 'backbone.conv_encoder.model.embedder.embedder.convolution.weight') ) rename_keys.append(('backbone.0.body.bn1.weight', 'backbone.conv_encoder.model.embedder.embedder.normalization.weight') ) rename_keys.append(('backbone.0.body.bn1.bias', 'backbone.conv_encoder.model.embedder.embedder.normalization.bias') ) rename_keys.append(('backbone.0.body.bn1.running_mean', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_mean') ) rename_keys.append(('backbone.0.body.bn1.running_var', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_var') ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var''', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var''', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ] ) return rename_keys def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : str = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> str: _lowercase : Any = '' if is_panoptic: _lowercase : Optional[Any] = 'detr.' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowercase : int = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Tuple = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[str] = in_proj_weight[:256, :] _lowercase : Tuple = in_proj_bias[:256] _lowercase : List[Any] = in_proj_weight[256:512, :] _lowercase : Any = in_proj_bias[256:512] _lowercase : int = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _lowercase : str = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Optional[int] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : Union[str, Any] = in_proj_weight[:256, :] _lowercase : Dict = in_proj_bias[:256] _lowercase : Tuple = in_proj_weight[256:512, :] _lowercase : Dict = in_proj_bias[256:512] _lowercase : str = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _lowercase : Tuple = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) _lowercase : Dict = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _lowercase : List[str] = in_proj_weight_cross_attn[:256, :] _lowercase : Tuple = in_proj_bias_cross_attn[:256] _lowercase : str = in_proj_weight_cross_attn[256:512, :] _lowercase : Union[str, Any] = in_proj_bias_cross_attn[256:512] _lowercase : List[Any] = in_proj_weight_cross_attn[-256:, :] _lowercase : Dict = in_proj_bias_cross_attn[-256:] def UpperCamelCase_( ) -> List[Any]: _lowercase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : str = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) return im @torch.no_grad() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=False ) -> List[Any]: _lowercase , _lowercase : int = get_detr_config(lowerCamelCase_ ) # load original model from torch hub _lowercase : int = { 'detr-resnet-50': 'detr_resnet50', 'detr-resnet-101': 'detr_resnet101', } logger.info(F'''Converting model {model_name}...''' ) _lowercase : Optional[Any] = torch.hub.load('facebookresearch/detr' , model_name_to_original_name[model_name] , pretrained=lowerCamelCase_ ).eval() _lowercase : str = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCamelCase_ ): if is_panoptic: _lowercase : str = 'detr.' + src rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCamelCase_ , is_panoptic=lowerCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowercase : List[Any] = 'detr.model.' if is_panoptic else 'model.' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('detr' ) and not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ) ): _lowercase : Tuple = state_dict.pop(lowerCamelCase_ ) _lowercase : int = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val elif key.startswith('bbox_attention' ) or key.startswith('mask_head' ): continue else: _lowercase : Optional[Any] = state_dict.pop(lowerCamelCase_ ) _lowercase : Union[str, Any] = val else: if not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : List[str] = val # finally, create HuggingFace model and load state dict _lowercase : Optional[Any] = DetrForSegmentation(lowerCamelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCamelCase_ ) model.load_state_dict(lowerCamelCase_ ) model.eval() # verify our conversion on an image _lowercase : str = 'coco_panoptic' if is_panoptic else 'coco_detection' _lowercase : Optional[int] = DetrImageProcessor(format=lowerCamelCase_ ) _lowercase : str = processor(images=prepare_img() , return_tensors='pt' ) _lowercase : Tuple = encoding['pixel_values'] _lowercase : int = detr(lowerCamelCase_ ) _lowercase : Tuple = model(lowerCamelCase_ ) assert torch.allclose(outputs.logits , original_outputs['pred_logits'] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs['pred_boxes'] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['pred_masks'] , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info('Uploading PyTorch model and image processor to the hub...' ) model.push_to_hub(F'''nielsr/{model_name}''' ) processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_name", default="detr-resnet-50", type=str, choices=["detr-resnet-50", "detr-resnet-101"], help="Name of the DETR model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the model to the hub or not.") SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCamelCase : Optional[int] = logging.get_logger(__name__) lowerCamelCase : Union[str, Any] = {"vocab_file": "vocab.txt", "emoji_file": "emoji.json"} lowerCamelCase : Tuple = { "vocab_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt", }, "emoji_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json", }, } lowerCamelCase : Any = { "abeja/gpt-neox-japanese-2.7b": 2_048, } def _SCREAMING_SNAKE_CASE ( lowercase : Tuple , lowercase : Any ): '''simple docstring''' with open(lowercase , 'r' , encoding='utf-8' ) as f: lowerCamelCase_ = json.loads(f.read() ) lowerCamelCase_ = collections.OrderedDict() lowerCamelCase_ = collections.OrderedDict() lowerCamelCase_ = collections.OrderedDict() with open(lowercase , 'r' , encoding='utf-8' ) as f: lowerCamelCase_ = f.readlines() lowerCamelCase_ = [[t.rstrip('\n' )] if (t == ',' or ',' not in t) else t.rstrip('\n' ).split(',' ) for t in token] for idx, b in enumerate(lowercase ): lowerCamelCase_ = b lowerCamelCase_ = idx for wd in b: lowerCamelCase_ = idx return vocab, raw_vocab, ids_to_tokens, emoji class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = ['''input_ids''', '''attention_mask'''] def __init__( self : str , A_ : Any , A_ : Any , A_ : Optional[Any]="<|endoftext|>" , A_ : Any="<|endoftext|>" , A_ : Optional[int]="<|startoftext|>" , A_ : Union[str, Any]="<|endoftext|>" , A_ : Any=False , **A_ : Tuple , ) -> Dict: """simple docstring""" super().__init__( unk_token=A_ , pad_token=A_ , bos_token=A_ , eos_token=A_ , do_clean_text=A_ , **A_ , ) if not os.path.isfile(A_ ): raise ValueError( f"""Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained""" ' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) if not os.path.isfile(A_ ): raise ValueError( f"""Can't find a emoji file at path '{emoji_file}'. To load the emoji information from a Google""" ' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`' ) lowerCamelCase_ = do_clean_text lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = load_vocab_and_emoji(A_ , A_ ) lowerCamelCase_ = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def a__ ( self : List[str] ) -> Dict: """simple docstring""" return len(self.raw_vocab ) def a__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" return dict(self.raw_vocab , **self.added_tokens_encoder ) def a__ ( self : Optional[Any] , A_ : str ) -> Tuple: """simple docstring""" return self.subword_tokenizer.tokenize(A_ , clean=self.do_clean_text ) def a__ ( self : Optional[int] , A_ : Dict ) -> List[Any]: """simple docstring""" return self.vocab.get(A_ , self.vocab.get(self.unk_token ) ) def a__ ( self : Union[str, Any] , A_ : Union[str, Any] ) -> int: """simple docstring""" return self.subword_tokenizer.convert_id_to_token(A_ ) def a__ ( self : Optional[int] , A_ : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = ''.join(A_ ).strip() return out_string def a__ ( self : Optional[Any] , A_ : "Conversation" ) -> List[int]: """simple docstring""" lowerCamelCase_ = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(A_ , add_special_tokens=A_ ) + [self.eos_token_id] ) if len(A_ ) > self.model_max_length: lowerCamelCase_ = input_ids[-self.model_max_length :] return input_ids def a__ ( self : List[Any] , A_ : str , A_ : Optional[str] = None ) -> Tuple[str]: """simple docstring""" lowerCamelCase_ = 0 if os.path.isdir(A_ ): lowerCamelCase_ = os.path.join( A_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowerCamelCase_ = os.path.join( A_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['emoji_file'] ) else: lowerCamelCase_ = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['vocab_file'] ) lowerCamelCase_ = ( (filename_prefix + '-' if filename_prefix else '') + save_directory + VOCAB_FILES_NAMES['emoji_file'] ) with open(A_ , 'w' , encoding='utf-8' ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ' Please check that the vocabulary is not corrupted!' ) lowerCamelCase_ = token_index writer.write(','.join(A_ ) + '\n' ) index += 1 with open(A_ , 'w' , encoding='utf-8' ) as writer: json.dump(self.emoji , A_ ) return vocab_file, emoji_file class A( UpperCamelCase ): '''simple docstring''' def __init__( self : Any , A_ : Union[str, Any] , A_ : int , A_ : Tuple ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = vocab # same as swe lowerCamelCase_ = ids_to_tokens # same as bpe lowerCamelCase_ = emoji lowerCamelCase_ = np.max([len(A_ ) for w in self.vocab.keys()] ) lowerCamelCase_ = re.compile(r'(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)' ) lowerCamelCase_ = re.compile(r'[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*' ) lowerCamelCase_ = re.compile(r'[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}' ) lowerCamelCase_ = re.compile( r'([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) lowerCamelCase_ = re.compile( r'(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*' ) lowerCamelCase_ = re.compile( r'((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*' ) lowerCamelCase_ = '─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿' lowerCamelCase_ = '▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟' lowerCamelCase_ = str.maketrans({k: '<BLOCK>' for k in keisen + blocks} ) def __len__( self : str ) -> Optional[int]: """simple docstring""" return len(self.ids_to_tokens ) def a__ ( self : Union[str, Any] , A_ : Dict ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = self.content_repattera.sub('<URL>' , A_ ) lowerCamelCase_ = self.content_repattera.sub('<EMAIL>' , A_ ) lowerCamelCase_ = self.content_repattera.sub('<TEL>' , A_ ) lowerCamelCase_ = self.content_repattera.sub('<DATE>' , A_ ) lowerCamelCase_ = self.content_repattera.sub('<DATE>' , A_ ) lowerCamelCase_ = self.content_repattera.sub('<PRICE>' , A_ ) lowerCamelCase_ = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: lowerCamelCase_ = content.replace('<BLOCK><BLOCK>' , '<BLOCK>' ) return content def a__ ( self : int , A_ : Optional[Any] , A_ : Tuple=False ) -> Dict: """simple docstring""" lowerCamelCase_ = text.replace(' ' , '<SP>' ) lowerCamelCase_ = text.replace(' ' , '<SP>' ) lowerCamelCase_ = text.replace('\r\n' , '<BR>' ) lowerCamelCase_ = text.replace('\n' , '<BR>' ) lowerCamelCase_ = text.replace('\r' , '<BR>' ) lowerCamelCase_ = text.replace('\t' , '<TAB>' ) lowerCamelCase_ = text.replace('—' , 'ー' ) lowerCamelCase_ = text.replace('−' , 'ー' ) for k, v in self.emoji["emoji"].items(): if k in text: lowerCamelCase_ = text.replace(A_ , A_ ) if clean: lowerCamelCase_ = self.clean_text(A_ ) def check_simbol(A_ : Union[str, Any] ): lowerCamelCase_ = x.encode() if len(A_ ) == 1 and len(A_ ) == 2: lowerCamelCase_ = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0XC2_A1 and c <= 0XC2_BF) or (c >= 0XC7_80 and c <= 0XC7_83) or (c >= 0XCA_B9 and c <= 0XCB_BF) or (c >= 0XCC_80 and c <= 0XCD_A2) ): return True return False def checkuae(A_ : Tuple ): lowerCamelCase_ = x.encode() if len(A_ ) == 1 and len(A_ ) == 3: lowerCamelCase_ = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0XE2_80_80 and c <= 0XE2_B0_7F: return True return False lowerCamelCase_ = 0 lowerCamelCase_ = [] while pos < len(A_ ): lowerCamelCase_ = min(len(A_ ) , pos + self.maxlen + 1 ) if text[pos] == '<' else pos + 3 lowerCamelCase_ = [] # (token_id, token, pos) for e in range(A_ , A_ , -1 ): lowerCamelCase_ = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(A_ ) > 2: lowerCamelCase_ = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(A_ ) > 0: # the smallest token_id is adopted lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = sorted(A_ , key=lambda A_ : x[0] )[0] result.append(A_ ) lowerCamelCase_ = e else: lowerCamelCase_ = pos + 1 lowerCamelCase_ = text[pos:end] if check_simbol(A_ ): result.append('<KIGOU>' ) elif checkuae(A_ ): result.append('<U2000U2BFF>' ) else: for i in wd.encode('utf-8' ): result.append('<|byte%d|>' % i ) lowerCamelCase_ = end return result def a__ ( self : List[Any] , A_ : Tuple , A_ : List[str]="\n" ) -> List[str]: """simple docstring""" lowerCamelCase_ = [] lowerCamelCase_ = [] lowerCamelCase_ = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(A_ ) > 0: words.append(bytearray(A_ ).decode('utf-8' , errors='replace' ) ) lowerCamelCase_ = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['emoji_inv'][word] ) elif word == "<SP>": words.append(' ' ) elif word == "<BR>": words.append(A_ ) elif word == "<TAB>": words.append('\t' ) elif word == "<BLOCK>": words.append('▀' ) elif word == "<KIGOU>": words.append('ǀ' ) elif word == "<U2000U2BFF>": words.append('‖' ) else: words.append(A_ ) if len(A_ ) > 0: words.append(bytearray(A_ ).decode('utf-8' , errors='replace' ) ) lowerCamelCase_ = ''.join(A_ ) return text
70
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput SCREAMING_SNAKE_CASE : str = "scheduler_config.json" class _lowerCamelCase( _a ): lowercase_ : Any = 1 lowercase_ : Dict = 2 lowercase_ : Union[str, Any] = 3 lowercase_ : Tuple = 4 lowercase_ : Optional[Any] = 5 @dataclass class _lowerCamelCase( _a ): lowercase_ : jnp.ndarray class _lowerCamelCase: lowercase_ : Union[str, Any] = SCHEDULER_CONFIG_NAME lowercase_ : str = ["""dtype"""] lowercase_ : Dict = [] lowercase_ : int = True @classmethod def UpperCamelCase ( cls, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase=False, **lowerCamelCase, ) -> Optional[int]: """simple docstring""" _lowercase , _lowercase : Optional[int] = cls.load_config( pretrained_model_name_or_path=lowerCamelCase, subfolder=lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase, ) _lowercase , _lowercase : Tuple = cls.from_config(lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase) if hasattr(lowerCamelCase, 'create_state') and getattr(lowerCamelCase, 'has_state', lowerCamelCase): _lowercase : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, **lowerCamelCase) -> Any: """simple docstring""" self.save_config(save_directory=lowerCamelCase, push_to_hub=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return self._get_compatibles() @classmethod def UpperCamelCase ( cls) -> Any: """simple docstring""" _lowercase : Any = list(set([cls.__name__] + cls._compatibles)) _lowercase : Dict = importlib.import_module(__name__.split('.')[0]) _lowercase : Any = [ getattr(lowerCamelCase, lowerCamelCase) for c in compatible_classes_str if hasattr(lowerCamelCase, lowerCamelCase) ] return compatible_classes def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> jnp.ndarray: assert len(lowerCamelCase_ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(lowerCamelCase_ ) - x.ndim) ) , lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=0.9_99 , lowerCamelCase_=jnp.floataa ) -> jnp.ndarray: def alpha_bar(lowerCamelCase_ ): return math.cos((time_step + 0.0_08) / 1.0_08 * math.pi / 2 ) ** 2 _lowercase : List[Any] = [] for i in range(lowerCamelCase_ ): _lowercase : Any = i / num_diffusion_timesteps _lowercase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(lowerCamelCase_ ) / alpha_bar(lowerCamelCase_ ) , lowerCamelCase_ ) ) return jnp.array(lowerCamelCase_ , dtype=lowerCamelCase_ ) @flax.struct.dataclass class _lowerCamelCase: lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray @classmethod def UpperCamelCase ( cls, lowerCamelCase) -> str: """simple docstring""" _lowercase : int = scheduler.config if config.trained_betas is not None: _lowercase : str = jnp.asarray(config.trained_betas, dtype=scheduler.dtype) elif config.beta_schedule == "linear": _lowercase : List[Any] = jnp.linspace(config.beta_start, config.beta_end, config.num_train_timesteps, dtype=scheduler.dtype) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _lowercase : Dict = ( jnp.linspace( config.beta_start**0.5, config.beta_end**0.5, config.num_train_timesteps, dtype=scheduler.dtype) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _lowercase : Optional[int] = betas_for_alpha_bar(config.num_train_timesteps, dtype=scheduler.dtype) else: raise NotImplementedError( F'''beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}''') _lowercase : List[str] = 1.0 - betas _lowercase : Union[str, Any] = jnp.cumprod(lowerCamelCase, axis=0) return cls( alphas=lowerCamelCase, betas=lowerCamelCase, alphas_cumprod=lowerCamelCase, ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : str = state.alphas_cumprod _lowercase : str = alphas_cumprod[timesteps] ** 0.5 _lowercase : Optional[Any] = sqrt_alpha_prod.flatten() _lowercase : Tuple = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) _lowercase : Union[str, Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowercase : Optional[Any] = sqrt_one_minus_alpha_prod.flatten() _lowercase : int = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase , _lowercase : Optional[int] = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: _lowercase , _lowercase : Tuple = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
89
0
'''simple docstring''' from __future__ import annotations def a__ ( _SCREAMING_SNAKE_CASE : list , _SCREAMING_SNAKE_CASE : int | None = None , _SCREAMING_SNAKE_CASE : int | None = None ) -> None: """simple docstring""" if start is None: UpperCAmelCase_ : Union[str, Any] = 0 if end is None: UpperCAmelCase_ : List[str] = len(_SCREAMING_SNAKE_CASE ) - 1 if start >= end: return UpperCAmelCase_ : Optional[Any] = (start + end) // 2 slowsort(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) slowsort(_SCREAMING_SNAKE_CASE , mid + 1 , _SCREAMING_SNAKE_CASE ) if sequence[end] < sequence[mid]: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = sequence[mid], sequence[end] slowsort(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
71
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> float: if not nums: raise ValueError('List is empty' ) return sum(lowerCamelCase_ ) / len(lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
89
0
'''simple docstring''' _UpperCAmelCase : dict[str, float] = { "joule": 1.0, "kilojoule": 10_00, "megajoule": 1_00_00_00, "gigajoule": 10_00_00_00_00, "wattsecond": 1.0, "watthour": 36_00, "kilowatthour": 3_60_00_00, "newtonmeter": 1.0, "calorie_nutr": 41_86.8, "kilocalorie_nutr": 4_18_68_00.00, "electronvolt": 1.6_02_17_66_34e-19, "britishthermalunit_it": 10_55.0_55_85, "footpound": 1.35_5818, } def UpperCamelCase ( lowercase_ : str , lowercase_ : str , lowercase_ : float ) -> float: '''simple docstring''' if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowercase =( f'Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n' f'Valid values are: {", ".join(lowercase_ )}' ) raise ValueError(lowercase_ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
72
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def UpperCamelCase_( ) -> List[Any]: _lowercase : int = ArgumentParser('Transformers CLI tool' , usage='transformers-cli <command> [<args>]' ) _lowercase : Optional[Any] = parser.add_subparsers(help='transformers-cli command helpers' ) # Register commands ConvertCommand.register_subcommand(lowerCamelCase_ ) DownloadCommand.register_subcommand(lowerCamelCase_ ) EnvironmentCommand.register_subcommand(lowerCamelCase_ ) RunCommand.register_subcommand(lowerCamelCase_ ) ServeCommand.register_subcommand(lowerCamelCase_ ) UserCommands.register_subcommand(lowerCamelCase_ ) AddNewModelCommand.register_subcommand(lowerCamelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCamelCase_ ) LfsCommands.register_subcommand(lowerCamelCase_ ) PTtoTFCommand.register_subcommand(lowerCamelCase_ ) # Let's go _lowercase : Any = parser.parse_args() if not hasattr(lowerCamelCase_ , 'func' ): parser.print_help() exit(1 ) # Run _lowercase : Optional[int] = args.func(lowerCamelCase_ ) service.run() if __name__ == "__main__": main()
89
0
import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _snake_case ( A__ , unittest.TestCase ): _lowercase : Optional[int] = AudioLDMPipeline _lowercase : Union[str, Any] = TEXT_TO_AUDIO_PARAMS _lowercase : List[str] = TEXT_TO_AUDIO_BATCH_PARAMS _lowercase : Tuple = frozenset( [ '''num_inference_steps''', '''num_waveforms_per_prompt''', '''generator''', '''latents''', '''output_type''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def SCREAMING_SNAKE_CASE__ ( self) -> Dict: torch.manual_seed(0) SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=(32, 64) , class_embed_type='simple_projection' , projection_class_embeddings_input_dim=32 , class_embeddings_concat=a , ) SCREAMING_SNAKE_CASE = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , clip_sample=a , set_alpha_to_one=a , ) torch.manual_seed(0) SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0) SCREAMING_SNAKE_CASE = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , projection_dim=32 , ) SCREAMING_SNAKE_CASE = ClapTextModelWithProjection(a) SCREAMING_SNAKE_CASE = RobertaTokenizer.from_pretrained('hf-internal-testing/tiny-random-roberta' , model_max_length=77) SCREAMING_SNAKE_CASE = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=1_6000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=a , ) SCREAMING_SNAKE_CASE = SpeechTaHifiGan(a) SCREAMING_SNAKE_CASE = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'vocoder': vocoder, } return components def SCREAMING_SNAKE_CASE__ ( self , a , a=0) -> int: if str(a).startswith('mps'): SCREAMING_SNAKE_CASE = torch.manual_seed(a) else: SCREAMING_SNAKE_CASE = torch.Generator(device=a).manual_seed(a) SCREAMING_SNAKE_CASE = { 'prompt': 'A hammer hitting a wooden surface', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, } return inputs def SCREAMING_SNAKE_CASE__ ( self) -> Optional[Any]: SCREAMING_SNAKE_CASE = 'cpu' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = audioldm_pipe(**a) SCREAMING_SNAKE_CASE = output.audios[0] assert audio.ndim == 1 assert len(a) == 256 SCREAMING_SNAKE_CASE = audio[:10] SCREAMING_SNAKE_CASE = np.array( [-0.00_50, 0.00_50, -0.00_60, 0.00_33, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_33]) assert np.abs(audio_slice - expected_slice).max() < 1E-2 def SCREAMING_SNAKE_CASE__ ( self) -> Tuple: SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = 3 * [inputs['prompt']] # forward SCREAMING_SNAKE_CASE = audioldm_pipe(**a) SCREAMING_SNAKE_CASE = output.audios[0] SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = 3 * [inputs.pop('prompt')] SCREAMING_SNAKE_CASE = audioldm_pipe.tokenizer( a , padding='max_length' , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=a , return_tensors='pt' , ) SCREAMING_SNAKE_CASE = text_inputs['input_ids'].to(a) SCREAMING_SNAKE_CASE = audioldm_pipe.text_encoder( a , ) SCREAMING_SNAKE_CASE = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state SCREAMING_SNAKE_CASE = F.normalize(a , dim=-1) SCREAMING_SNAKE_CASE = prompt_embeds # forward SCREAMING_SNAKE_CASE = audioldm_pipe(**a) SCREAMING_SNAKE_CASE = output.audios[0] assert np.abs(audio_a - audio_a).max() < 1E-2 def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = 3 * ['this is a negative prompt'] SCREAMING_SNAKE_CASE = negative_prompt SCREAMING_SNAKE_CASE = 3 * [inputs['prompt']] # forward SCREAMING_SNAKE_CASE = audioldm_pipe(**a) SCREAMING_SNAKE_CASE = output.audios[0] SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = 3 * [inputs.pop('prompt')] SCREAMING_SNAKE_CASE = [] for p in [prompt, negative_prompt]: SCREAMING_SNAKE_CASE = audioldm_pipe.tokenizer( a , padding='max_length' , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=a , return_tensors='pt' , ) SCREAMING_SNAKE_CASE = text_inputs['input_ids'].to(a) SCREAMING_SNAKE_CASE = audioldm_pipe.text_encoder( a , ) SCREAMING_SNAKE_CASE = text_embeds.text_embeds # additional L_2 normalization over each hidden-state SCREAMING_SNAKE_CASE = F.normalize(a , dim=-1) embeds.append(a) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = embeds # forward SCREAMING_SNAKE_CASE = audioldm_pipe(**a) SCREAMING_SNAKE_CASE = output.audios[0] assert np.abs(audio_a - audio_a).max() < 1E-2 def SCREAMING_SNAKE_CASE__ ( self) -> Dict: SCREAMING_SNAKE_CASE = 'cpu' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = PNDMScheduler(skip_prk_steps=a) SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = 'egg cracking' SCREAMING_SNAKE_CASE = audioldm_pipe(**a , negative_prompt=a) SCREAMING_SNAKE_CASE = output.audios[0] assert audio.ndim == 1 assert len(a) == 256 SCREAMING_SNAKE_CASE = audio[:10] SCREAMING_SNAKE_CASE = np.array( [-0.00_51, 0.00_50, -0.00_60, 0.00_34, -0.00_26, 0.00_33, -0.00_27, 0.00_33, -0.00_28, 0.00_32]) assert np.abs(audio_slice - expected_slice).max() < 1E-2 def SCREAMING_SNAKE_CASE__ ( self) -> List[Any]: SCREAMING_SNAKE_CASE = 'cpu' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = PNDMScheduler(skip_prk_steps=a) SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = 'A hammer hitting a wooden surface' # test num_waveforms_per_prompt=1 (default) SCREAMING_SNAKE_CASE = audioldm_pipe(a , num_inference_steps=2).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = audioldm_pipe([prompt] * batch_size , num_inference_steps=2).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = audioldm_pipe(a , num_inference_steps=2 , num_waveforms_per_prompt=a).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=a).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def SCREAMING_SNAKE_CASE__ ( self) -> str: SCREAMING_SNAKE_CASE = 'cpu' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = audioldm_pipe.vocoder.config.sampling_rate SCREAMING_SNAKE_CASE = self.get_dummy_inputs(a) SCREAMING_SNAKE_CASE = audioldm_pipe(audio_length_in_s=0.0_16 , **a) SCREAMING_SNAKE_CASE = output.audios[0] assert audio.ndim == 1 assert len(a) / vocoder_sampling_rate == 0.0_16 SCREAMING_SNAKE_CASE = audioldm_pipe(audio_length_in_s=0.0_32 , **a) SCREAMING_SNAKE_CASE = output.audios[0] assert audio.ndim == 1 assert len(a) / vocoder_sampling_rate == 0.0_32 def SCREAMING_SNAKE_CASE__ ( self) -> Optional[int]: SCREAMING_SNAKE_CASE = self.get_dummy_components() SCREAMING_SNAKE_CASE = AudioLDMPipeline(**a) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = ['hey'] SCREAMING_SNAKE_CASE = audioldm_pipe(a , num_inference_steps=1) SCREAMING_SNAKE_CASE = output.audios.shape assert audio_shape == (1, 256) SCREAMING_SNAKE_CASE = audioldm_pipe.vocoder.config config.model_in_dim *= 2 SCREAMING_SNAKE_CASE = SpeechTaHifiGan(a).to(a) SCREAMING_SNAKE_CASE = audioldm_pipe(a , num_inference_steps=1) SCREAMING_SNAKE_CASE = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def SCREAMING_SNAKE_CASE__ ( self) -> Dict: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=a) def SCREAMING_SNAKE_CASE__ ( self) -> Optional[Any]: self._test_inference_batch_single_identical(test_mean_pixel_difference=a) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def SCREAMING_SNAKE_CASE__ ( self) -> int: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=a) @slow class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self , a , a="cpu" , a=torch.floataa , a=0) -> Dict: SCREAMING_SNAKE_CASE = torch.Generator(device=a).manual_seed(a) SCREAMING_SNAKE_CASE = np.random.RandomState(a).standard_normal((1, 8, 128, 16)) SCREAMING_SNAKE_CASE = torch.from_numpy(a).to(device=a , dtype=a) SCREAMING_SNAKE_CASE = { 'prompt': 'A hammer hitting a wooden surface', 'latents': latents, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 2.5, } return inputs def SCREAMING_SNAKE_CASE__ ( self) -> Any: SCREAMING_SNAKE_CASE = AudioLDMPipeline.from_pretrained('cvssp/audioldm') SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = self.get_inputs(a) SCREAMING_SNAKE_CASE = 25 SCREAMING_SNAKE_CASE = audioldm_pipe(**a).audios[0] assert audio.ndim == 1 assert len(a) == 8_1920 SCREAMING_SNAKE_CASE = audio[7_7230:7_7240] SCREAMING_SNAKE_CASE = np.array( [-0.48_84, -0.46_07, 0.00_23, 0.50_07, 0.58_96, 0.51_51, 0.38_13, -0.02_08, -0.36_87, -0.43_15]) SCREAMING_SNAKE_CASE = np.abs(expected_slice - audio_slice).max() assert max_diff < 1E-2 def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: SCREAMING_SNAKE_CASE = AudioLDMPipeline.from_pretrained('cvssp/audioldm') SCREAMING_SNAKE_CASE = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config) SCREAMING_SNAKE_CASE = audioldm_pipe.to(a) audioldm_pipe.set_progress_bar_config(disable=a) SCREAMING_SNAKE_CASE = self.get_inputs(a) SCREAMING_SNAKE_CASE = audioldm_pipe(**a).audios[0] assert audio.ndim == 1 assert len(a) == 8_1920 SCREAMING_SNAKE_CASE = audio[2_7780:2_7790] SCREAMING_SNAKE_CASE = np.array([-0.21_31, -0.08_73, -0.01_24, -0.01_89, 0.05_69, 0.13_73, 0.18_83, 0.28_86, 0.32_97, 0.22_12]) SCREAMING_SNAKE_CASE = np.abs(expected_slice - audio_slice).max() assert max_diff < 3E-2
73
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Optional[int] = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def UpperCamelCase ( self, lowerCamelCase=0) -> str: """simple docstring""" _lowercase : Optional[int] = np.random.RandomState(lowerCamelCase) _lowercase : Union[str, Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : List[Any] = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[str] = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Any = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_dummy_inputs() _lowercase : Any = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : Optional[int] = output.images[0, -3:, -3:, -1] _lowercase : int = self.get_dummy_inputs() _lowercase : Union[str, Any] = 3 * [inputs.pop('prompt')] _lowercase : Union[str, Any] = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Tuple = text_inputs['input_ids'] _lowercase : Any = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0] _lowercase : List[Any] = prompt_embeds # forward _lowercase : Union[str, Any] = pipe(**lowerCamelCase) _lowercase : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs() _lowercase : Any = 3 * ['this is a negative prompt'] _lowercase : str = negative_prompt _lowercase : Optional[int] = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : str = output.images[0, -3:, -3:, -1] _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : str = 3 * [inputs.pop('prompt')] _lowercase : Optional[int] = [] for p in [prompt, negative_prompt]: _lowercase : Tuple = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Dict = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0]) _lowercase , _lowercase : str = embeds # forward _lowercase : Dict = pipe(**lowerCamelCase) _lowercase : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = ort.SessionOptions() _lowercase : str = False return options def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'A painting of a squirrel eating a burger' np.random.seed(0) _lowercase : Union[str, Any] = sd_pipe([prompt], guidance_scale=6.0, num_inference_steps=10, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : str = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'open neural network exchange' _lowercase : List[Any] = np.random.RandomState(0) _lowercase : Optional[Any] = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = 'open neural network exchange' _lowercase : str = np.random.RandomState(0) _lowercase : Dict = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[Any] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : List[Any] = 0 def test_callback_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase) -> None: _lowercase : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _lowercase : Any = latents[0, -3:, -3:, -1] _lowercase : Tuple = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _lowercase : List[Any] = latents[0, -3:, -3:, -1] _lowercase : str = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 _lowercase : Any = False _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = 'Andromeda galaxy in a bottle' _lowercase : str = np.random.RandomState(0) pipe( prompt=lowerCamelCase, num_inference_steps=5, guidance_scale=7.5, generator=lowerCamelCase, callback=lowerCamelCase, callback_steps=1, ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) assert isinstance(lowerCamelCase, lowerCamelCase) assert pipe.safety_checker is None _lowercase : Optional[int] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase) _lowercase : Any = OnnxStableDiffusionPipeline.from_pretrained(lowerCamelCase) # sanity check that the pipeline still works assert pipe.safety_checker is None _lowercase : List[str] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None
89
0
import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """google/owlvit-base-patch32""": """https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json""", """google/owlvit-base-patch16""": """https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json""", """google/owlvit-large-patch14""": """https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json""", } class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''owlvit_text_model''' def __init__( self : Dict , _A : int=4_9408 , _A : int=512 , _A : Optional[int]=2048 , _A : List[str]=12 , _A : Any=8 , _A : str=16 , _A : Union[str, Any]="quick_gelu" , _A : Optional[Any]=1e-5 , _A : Tuple=0.0 , _A : Tuple=0.02 , _A : Tuple=1.0 , _A : Union[str, Any]=0 , _A : Tuple=4_9406 , _A : Any=4_9407 , **_A : Tuple , ): """simple docstring""" super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A ) __SCREAMING_SNAKE_CASE : Any = vocab_size __SCREAMING_SNAKE_CASE : Optional[Any] = hidden_size __SCREAMING_SNAKE_CASE : Dict = intermediate_size __SCREAMING_SNAKE_CASE : int = num_hidden_layers __SCREAMING_SNAKE_CASE : List[str] = num_attention_heads __SCREAMING_SNAKE_CASE : List[str] = max_position_embeddings __SCREAMING_SNAKE_CASE : int = hidden_act __SCREAMING_SNAKE_CASE : int = layer_norm_eps __SCREAMING_SNAKE_CASE : Tuple = attention_dropout __SCREAMING_SNAKE_CASE : Union[str, Any] = initializer_range __SCREAMING_SNAKE_CASE : Tuple = initializer_factor @classmethod def UpperCAmelCase__ ( cls : Optional[int] , _A : Union[str, os.PathLike] , **_A : Tuple ): """simple docstring""" cls._set_token_in_kwargs(_A ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : List[str] = cls.get_config_dict(_A , **_A ) # get the text config dict if we are loading from OwlViTConfig if config_dict.get('''model_type''' ) == "owlvit": __SCREAMING_SNAKE_CASE : Optional[int] = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_A , **_A ) class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''owlvit_vision_model''' def __init__( self : List[str] , _A : Optional[Any]=768 , _A : List[Any]=3072 , _A : Union[str, Any]=12 , _A : Dict=12 , _A : Tuple=3 , _A : Any=768 , _A : Optional[Any]=32 , _A : List[str]="quick_gelu" , _A : Optional[int]=1e-5 , _A : Union[str, Any]=0.0 , _A : Tuple=0.02 , _A : Union[str, Any]=1.0 , **_A : Any , ): """simple docstring""" super().__init__(**_A ) __SCREAMING_SNAKE_CASE : int = hidden_size __SCREAMING_SNAKE_CASE : Tuple = intermediate_size __SCREAMING_SNAKE_CASE : List[str] = num_hidden_layers __SCREAMING_SNAKE_CASE : Tuple = num_attention_heads __SCREAMING_SNAKE_CASE : Optional[int] = num_channels __SCREAMING_SNAKE_CASE : List[str] = image_size __SCREAMING_SNAKE_CASE : Dict = patch_size __SCREAMING_SNAKE_CASE : Optional[int] = hidden_act __SCREAMING_SNAKE_CASE : Any = layer_norm_eps __SCREAMING_SNAKE_CASE : List[str] = attention_dropout __SCREAMING_SNAKE_CASE : Dict = initializer_range __SCREAMING_SNAKE_CASE : Optional[Any] = initializer_factor @classmethod def UpperCAmelCase__ ( cls : Union[str, Any] , _A : Union[str, os.PathLike] , **_A : Optional[Any] ): """simple docstring""" cls._set_token_in_kwargs(_A ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : int = cls.get_config_dict(_A , **_A ) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get('''model_type''' ) == "owlvit": __SCREAMING_SNAKE_CASE : Union[str, Any] = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_A , **_A ) class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''owlvit''' lowerCAmelCase_ = True def __init__( self : int , _A : Optional[Any]=None , _A : List[Any]=None , _A : Optional[Any]=512 , _A : List[str]=2.65_92 , _A : List[Any]=True , **_A : List[Any] , ): """simple docstring""" super().__init__(**_A ) if text_config is None: __SCREAMING_SNAKE_CASE : Optional[Any] = {} logger.info('''text_config is None. Initializing the OwlViTTextConfig with default values.''' ) if vision_config is None: __SCREAMING_SNAKE_CASE : Union[str, Any] = {} logger.info('''vision_config is None. initializing the OwlViTVisionConfig with default values.''' ) __SCREAMING_SNAKE_CASE : Any = OwlViTTextConfig(**_A ) __SCREAMING_SNAKE_CASE : Optional[Any] = OwlViTVisionConfig(**_A ) __SCREAMING_SNAKE_CASE : Tuple = projection_dim __SCREAMING_SNAKE_CASE : List[str] = logit_scale_init_value __SCREAMING_SNAKE_CASE : Any = return_dict __SCREAMING_SNAKE_CASE : Optional[Any] = 1.0 @classmethod def UpperCAmelCase__ ( cls : Union[str, Any] , _A : Union[str, os.PathLike] , **_A : Optional[Any] ): """simple docstring""" cls._set_token_in_kwargs(_A ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Tuple = cls.get_config_dict(_A , **_A ) if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_A , **_A ) @classmethod def UpperCAmelCase__ ( cls : Any , _A : Dict , _A : Dict , **_A : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = {} __SCREAMING_SNAKE_CASE : Union[str, Any] = text_config __SCREAMING_SNAKE_CASE : Optional[int] = vision_config return cls.from_dict(_A , **_A ) def UpperCAmelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = copy.deepcopy(self.__dict__ ) __SCREAMING_SNAKE_CASE : str = self.text_config.to_dict() __SCREAMING_SNAKE_CASE : Optional[int] = self.vision_config.to_dict() __SCREAMING_SNAKE_CASE : Any = self.__class__.model_type return output class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" @property def UpperCAmelCase__ ( self : Optional[Any] ): """simple docstring""" return OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''sequence'''}), ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''attention_mask''', {0: '''batch''', 1: '''sequence'''}), ] ) @property def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" return OrderedDict( [ ('''logits_per_image''', {0: '''batch'''}), ('''logits_per_text''', {0: '''batch'''}), ('''text_embeds''', {0: '''batch'''}), ('''image_embeds''', {0: '''batch'''}), ] ) @property def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" return 1e-4 def UpperCAmelCase__ ( self : List[str] , _A : "ProcessorMixin" , _A : int = -1 , _A : int = -1 , _A : Optional["TensorType"] = None , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = super().generate_dummy_inputs( processor.tokenizer , batch_size=_A , seq_length=_A , framework=_A ) __SCREAMING_SNAKE_CASE : Optional[int] = super().generate_dummy_inputs( processor.image_processor , batch_size=_A , framework=_A ) return {**text_input_dict, **image_input_dict} @property def UpperCAmelCase__ ( self : int ): """simple docstring""" return 14
74
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : List[Any] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = ["PoolFormerFeatureExtractor"] SCREAMING_SNAKE_CASE : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[Any] = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
89
0
'''simple docstring''' import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class lowerCamelCase_ : def __init__( self : Dict , _A : str=2 , _A : Tuple=3 , _A : Tuple=64 , _A : str=None ): '''simple docstring''' UpperCAmelCase__ : int = np.random.default_rng(_A ) UpperCAmelCase__ : Any = length UpperCAmelCase__ : int = rng.normal(size=(length,) ).astype(np.floataa ) UpperCAmelCase__ : Optional[Any] = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self : Tuple ): '''simple docstring''' return self.length def __getitem__( self : Union[str, Any] , _A : Optional[Any] ): '''simple docstring''' return {"x": self.x[i], "y": self.y[i]} class lowerCamelCase_ ( torch.nn.Module ): def __init__( self : str , _A : List[str]=0 , _A : List[Any]=0 , _A : int=False ): '''simple docstring''' super().__init__() UpperCAmelCase__ : Dict = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase__ : Dict = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCAmelCase__ : List[Any] = True def lowercase_ ( self : Any , _A : List[Any]=None ): '''simple docstring''' if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) UpperCAmelCase__ : Any = False return x * self.a[0] + self.b[0] class lowerCamelCase_ ( torch.nn.Module ): def __init__( self : Optional[int] , _A : List[str]=0 , _A : Optional[int]=0 , _A : Any=False ): '''simple docstring''' super().__init__() UpperCAmelCase__ : Dict = torch.nn.Parameter(torch.tensor(_A ).float() ) UpperCAmelCase__ : Optional[Any] = torch.nn.Parameter(torch.tensor(_A ).float() ) UpperCAmelCase__ : str = True def lowercase_ ( self : str , _A : List[Any]=None ): '''simple docstring''' if self.first_batch: print(f"""Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}""" ) UpperCAmelCase__ : int = False return x * self.a + self.b def a__ ( lowerCAmelCase__ , lowerCAmelCase__ = 16 ) -> Dict: from datasets import load_dataset from transformers import AutoTokenizer UpperCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) UpperCAmelCase__ : Dict = {'''train''': '''tests/test_samples/MRPC/train.csv''', '''validation''': '''tests/test_samples/MRPC/dev.csv'''} UpperCAmelCase__ : str = load_dataset('''csv''' , data_files=lowerCAmelCase__ ) UpperCAmelCase__ : Dict = datasets['''train'''].unique('''label''' ) UpperCAmelCase__ : Union[str, Any] = {v: i for i, v in enumerate(lowerCAmelCase__ )} def tokenize_function(lowerCAmelCase__ ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase__ : Any = tokenizer( examples['''sentence1'''] , examples['''sentence2'''] , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , padding='''max_length''' ) if "label" in examples: UpperCAmelCase__ : Union[str, Any] = [label_to_id[l] for l in examples['''label''']] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase__ : Optional[int] = datasets.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , remove_columns=['''sentence1''', '''sentence2''', '''label'''] , ) def collate_fn(lowerCAmelCase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowerCAmelCase__ , padding='''max_length''' , max_length=1_28 , return_tensors='''pt''' ) return tokenizer.pad(lowerCAmelCase__ , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. UpperCAmelCase__ : Optional[Any] = DataLoader(tokenized_datasets['''train'''] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=2 ) UpperCAmelCase__ : Tuple = DataLoader(tokenized_datasets['''validation'''] , shuffle=lowerCAmelCase__ , collate_fn=lowerCAmelCase__ , batch_size=1 ) return train_dataloader, eval_dataloader
75
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore SCREAMING_SNAKE_CASE : int = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" SCREAMING_SNAKE_CASE : Dict = [file for file in filepaths if file != file.lower()] if upper_files: print(F"{len(upper_files)} files contain uppercase characters:") print("\n".join(upper_files) + "\n") SCREAMING_SNAKE_CASE : List[Any] = [file for file in filepaths if " " in file] if space_files: print(F"{len(space_files)} files contain space characters:") print("\n".join(space_files) + "\n") SCREAMING_SNAKE_CASE : Any = [file for file in filepaths if "-" in file] if hyphen_files: print(F"{len(hyphen_files)} files contain hyphen characters:") print("\n".join(hyphen_files) + "\n") SCREAMING_SNAKE_CASE : str = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"{len(nodir_files)} files are not in a directory:") print("\n".join(nodir_files) + "\n") SCREAMING_SNAKE_CASE : Tuple = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
89
0
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'BAAI/AltCLIP': 'https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="altclip_text_model" def __init__( self , UpperCamelCase_=25_00_02 , UpperCamelCase_=10_24 , UpperCamelCase_=24 , UpperCamelCase_=16 , UpperCamelCase_=40_96 , UpperCamelCase_="gelu" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=5_14 , UpperCamelCase_=1 , UpperCamelCase_=0.0_2 , UpperCamelCase_=0.0_2 , UpperCamelCase_=1E-05 , UpperCamelCase_=1 , UpperCamelCase_=0 , UpperCamelCase_=2 , UpperCamelCase_="absolute" , UpperCamelCase_=True , UpperCamelCase_=7_68 , **UpperCamelCase_ , ) -> List[str]: super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowercase : List[str] = vocab_size __lowercase : Optional[int] = hidden_size __lowercase : Optional[Any] = num_hidden_layers __lowercase : Optional[int] = num_attention_heads __lowercase : Optional[Any] = hidden_act __lowercase : int = intermediate_size __lowercase : str = hidden_dropout_prob __lowercase : List[Any] = attention_probs_dropout_prob __lowercase : int = max_position_embeddings __lowercase : Dict = type_vocab_size __lowercase : int = initializer_range __lowercase : Optional[int] = initializer_factor __lowercase : Union[str, Any] = layer_norm_eps __lowercase : int = position_embedding_type __lowercase : Optional[Any] = use_cache __lowercase : Optional[Any] = project_dim class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="altclip_vision_model" def __init__( self , UpperCamelCase_=7_68 , UpperCamelCase_=30_72 , UpperCamelCase_=5_12 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=3 , UpperCamelCase_=2_24 , UpperCamelCase_=32 , UpperCamelCase_="quick_gelu" , UpperCamelCase_=1E-5 , UpperCamelCase_=0.0 , UpperCamelCase_=0.0_2 , UpperCamelCase_=1.0 , **UpperCamelCase_ , ) -> Optional[int]: super().__init__(**UpperCamelCase_ ) __lowercase : Tuple = hidden_size __lowercase : Optional[int] = intermediate_size __lowercase : int = projection_dim __lowercase : Tuple = num_hidden_layers __lowercase : str = num_attention_heads __lowercase : str = num_channels __lowercase : int = patch_size __lowercase : List[str] = image_size __lowercase : Optional[int] = initializer_range __lowercase : Union[str, Any] = initializer_factor __lowercase : Optional[int] = attention_dropout __lowercase : str = layer_norm_eps __lowercase : Dict = hidden_act @classmethod def _lowerCamelCase ( cls , UpperCamelCase_ , **UpperCamelCase_ ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCamelCase_ ) __lowercase ,__lowercase : Optional[Any] = cls.get_config_dict(UpperCamelCase_ , **UpperCamelCase_ ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('''model_type''' ) == "altclip": __lowercase : int = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(UpperCamelCase_ , **UpperCamelCase_ ) class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="altclip" UpperCamelCase =True def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=7_68 , UpperCamelCase_=2.6_5_9_2 , **UpperCamelCase_ ) -> Optional[int]: # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). __lowercase : Optional[Any] = kwargs.pop('''text_config_dict''' , UpperCamelCase_ ) __lowercase : Tuple = kwargs.pop('''vision_config_dict''' , UpperCamelCase_ ) super().__init__(**UpperCamelCase_ ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __lowercase : Any = {} # This is the complete result when using `text_config_dict`. __lowercase : Any = AltCLIPTextConfig(**UpperCamelCase_ ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __lowercase : Union[str, Any] = ( F"""`{key}` is found in both `text_config_dict` and `text_config` but with different values. """ F"""The value `text_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: __lowercase : Tuple = ( F"""`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The """ F"""value `text_config[\"{key}\"]` will be overriden.""" ) logger.warning(UpperCamelCase_ ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __lowercase : Optional[Any] = {} # This is the complete result when using `vision_config_dict`. __lowercase : List[str] = AltCLIPVisionConfig(**UpperCamelCase_ ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __lowercase : Any = { str(UpperCamelCase_ ): value for key, value in _vision_config_dict['''id2label'''].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __lowercase : Union[str, Any] = ( F"""`{key}` is found in both `vision_config_dict` and `vision_config` but with different """ F"""values. The value `vision_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: __lowercase : Optional[int] = ( F"""`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. """ F"""The value `vision_config[\"{key}\"]` will be overriden.""" ) logger.warning(UpperCamelCase_ ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __lowercase : Optional[int] = {} logger.info('''`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.''' ) if vision_config is None: __lowercase : Tuple = {} logger.info('''`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.''' ) __lowercase : Union[str, Any] = AltCLIPTextConfig(**UpperCamelCase_ ) __lowercase : List[Any] = AltCLIPVisionConfig(**UpperCamelCase_ ) __lowercase : str = projection_dim __lowercase : str = logit_scale_init_value __lowercase : List[Any] = 1.0 @classmethod def _lowerCamelCase ( cls , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) -> Optional[int]: return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **UpperCamelCase_ ) def _lowerCamelCase ( self ) -> Optional[int]: __lowercase : List[Any] = copy.deepcopy(self.__dict__ ) __lowercase : Optional[Any] = self.text_config.to_dict() __lowercase : int = self.vision_config.to_dict() __lowercase : Optional[int] = self.__class__.model_type return output
76
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Any: _lowercase : str = 10 _lowercase : List[str] = datasets.Features( { 'tokens': datasets.Sequence(datasets.Value('string' ) ), 'labels': datasets.Sequence(datasets.ClassLabel(names=['negative', 'positive'] ) ), 'answers': datasets.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), 'id': datasets.Value('int64' ), } ) _lowercase : Union[str, Any] = datasets.Dataset.from_dict( { 'tokens': [['foo'] * 5] * n, 'labels': [[1] * 5] * n, 'answers': [{'answer_start': [97], 'text': ['1976']}] * 10, 'id': list(range(lowerCamelCase_ ) ), } , features=lowerCamelCase_ , ) return dataset @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : int = str(tmp_path_factory.mktemp('data' ) / 'file.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return filename # FILE_CONTENT + files SCREAMING_SNAKE_CASE : str = "\\n Text data.\n Second line of data." @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'file.txt' _lowercase : List[str] = FILE_CONTENT with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: import bza _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.bz2' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'file.txt.gz' ) _lowercase : Optional[int] = bytes(lowerCamelCase_ , 'utf-8' ) with gzip.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: if datasets.config.LZ4_AVAILABLE: import lza.frame _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.lz4' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with lza.frame.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: if datasets.config.PY7ZR_AVAILABLE: import pyazr _lowercase : int = tmp_path_factory.mktemp('data' ) / 'file.txt.7z' with pyazr.SevenZipFile(lowerCamelCase_ , 'w' ) as archive: archive.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: import tarfile _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: import lzma _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.xz' _lowercase : int = bytes(lowerCamelCase_ , 'utf-8' ) with lzma.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: import zipfile _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'file.txt.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.zst' _lowercase : Dict = bytes(lowerCamelCase_ , 'utf-8' ) with zstd.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'file.xml' _lowercase : Optional[Any] = textwrap.dedent( '\\n <?xml version="1.0" encoding="UTF-8" ?>\n <tmx version="1.4">\n <header segtype="sentence" srclang="ca" />\n <body>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang="en"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang="en"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang="en"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang="en"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang="en"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>' ) with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] SCREAMING_SNAKE_CASE : Optional[Any] = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } SCREAMING_SNAKE_CASE : Tuple = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] SCREAMING_SNAKE_CASE : Any = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> List[str]: return DATA_DICT_OF_LISTS @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Optional[int] = datasets.Dataset.from_dict(lowerCamelCase_ ) _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.sqlite' ) with contextlib.closing(sqlitea.connect(lowerCamelCase_ ) ) as con: _lowercase : Union[str, Any] = con.cursor() cur.execute('CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)' ) for item in DATA: cur.execute('INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : Tuple = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset2.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : str = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: import bza _lowercase : int = tmp_path_factory.mktemp('data' ) / 'dataset.csv.bz2' with open(lowerCamelCase_ , 'rb' ) as f: _lowercase : int = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(csv_path.replace('.csv' , '.CSV' ) ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(csva_path.replace('.csv' , '.CSV' ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.parquet' ) _lowercase : Optional[Any] = pa.schema( { 'col_1': pa.string(), 'col_2': pa.intaa(), 'col_3': pa.floataa(), } ) with open(lowerCamelCase_ , 'wb' ) as f: _lowercase : List[str] = pq.ParquetWriter(lowerCamelCase_ , schema=lowerCamelCase_ ) _lowercase : Any = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(lowerCamelCase_ ) )] for k in DATA[0]} , schema=lowerCamelCase_ ) writer.write_table(lowerCamelCase_ ) writer.close() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : List[Any] = {'data': DATA} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : Optional[Any] = {'data': DATA_DICT_OF_LISTS} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Dict = str(tmp_path_factory.mktemp('data' ) / 'dataset2.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset_312.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_312: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset-str.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_STR: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: import gzip _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Dict: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Tuple = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Optional[int] = ['0', '1', '2', '3'] _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : str = ['0', '1', '2', '3'] _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : List[Any] = ['0', '1', '2', '3'] _lowercase : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset.abc' with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: _lowercase : List[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.ext.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported.ext' ) ) f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported_2.ext' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = '\n'.join(['First', 'Second\u2029with Unicode new line', 'Third'] ) _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset_with_unicode_new_lines.txt' ) with open(lowerCamelCase_ , 'w' , encoding='utf-8' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Dict: return os.path.join('tests' , 'features' , 'data' , 'test_image_rgb.jpg' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> int: return os.path.join('tests' , 'features' , 'data' , 'test_audio_44100.wav' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'dataset.img.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ).replace('.jpg' , '2.jpg' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data_dir' ) (data_dir / "subdir").mkdir() with open(data_dir / 'subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / 'subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden file with open(data_dir / 'subdir' / '.test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '.subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / '.subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) return data_dir
89
0
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> Optional[int]: """simple docstring""" return [ { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], }, { 0: [6], 1: [9], 2: [4, 5], 3: [4], 4: [2, 3], 5: [2], 6: [0, 7], 7: [6], 8: [], 9: [1], }, { 0: [4], 1: [6], 2: [], 3: [5, 6, 7], 4: [0, 6], 5: [3, 8, 9], 6: [1, 3, 4, 7], 7: [3, 6, 8, 9], 8: [5, 7], 9: [5, 7], }, { 0: [1, 3], 1: [0, 2, 4], 2: [1, 3, 4], 3: [0, 2, 4], 4: [1, 2, 3], }, ][index] def _UpperCamelCase ( UpperCamelCase ) -> list[tuple[int, int]]: """simple docstring""" __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : Any = len(UpperCamelCase ) # No of vertices in graph __UpperCAmelCase : Union[str, Any] = [0] * n __UpperCAmelCase : List[Any] = [False] * n def dfs(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = id_ id_ += 1 for to in graph[at]: if to == parent: pass elif not visited[to]: dfs(UpperCamelCase , UpperCamelCase , UpperCamelCase , id_ ) __UpperCAmelCase : Tuple = min(low[at] , low[to] ) if id_ <= low[to]: bridges.append((at, to) if at < to else (to, at) ) else: # This edge is a back edge and cannot be a bridge __UpperCAmelCase : List[Any] = min(low[at] , low[to] ) __UpperCAmelCase : list[tuple[int, int]] = [] for i in range(UpperCamelCase ): if not visited[i]: dfs(UpperCamelCase , -1 , UpperCamelCase , id_ ) return bridges if __name__ == "__main__": import doctest doctest.testmod()
77
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE : str = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : int = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer SCREAMING_SNAKE_CASE_: List[str] =logging.get_logger(__name__) SCREAMING_SNAKE_CASE_: Optional[int] ={'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all MVP models at https://huggingface.co/models?filter=mvp SCREAMING_SNAKE_CASE_: Any ={ 'vocab_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json', }, 'added_tokens.json': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json', }, 'merges_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt', }, 'tokenizer_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json', }, } SCREAMING_SNAKE_CASE_: int ={ 'RUCAIBox/mvp': 10_24, } class __A ( UpperCamelCase__ ): a__ : Optional[int] = VOCAB_FILES_NAMES a__ : str = PRETRAINED_VOCAB_FILES_MAP a__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : Any = ["""input_ids""", """attention_mask"""] a__ : List[Any] = MvpTokenizer def __init__(self : Dict , __a : List[Any]=None , __a : List[Any]=None , __a : Optional[int]=None , __a : Any="replace" , __a : Optional[Any]="<s>" , __a : List[str]="</s>" , __a : int="</s>" , __a : Optional[int]="<s>" , __a : str="<unk>" , __a : str="<pad>" , __a : List[Any]="<mask>" , __a : Tuple=False , __a : str=True , **__a : Optional[int] , ): super().__init__( __a , __a , tokenizer_file=__a , errors=__a , bos_token=__a , eos_token=__a , sep_token=__a , cls_token=__a , unk_token=__a , pad_token=__a , mask_token=__a , add_prefix_space=__a , trim_offsets=__a , **__a , ) UpperCAmelCase_ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , __a ) != add_prefix_space: UpperCAmelCase_ = getattr(__a , pre_tok_state.pop("type" ) ) UpperCAmelCase_ = add_prefix_space UpperCAmelCase_ = pre_tok_class(**__a ) UpperCAmelCase_ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` UpperCAmelCase_ = "post_processor" UpperCAmelCase_ = getattr(self.backend_tokenizer , __a , __a ) if tokenizer_component_instance: UpperCAmelCase_ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: UpperCAmelCase_ = tuple(state["sep"] ) if "cls" in state: UpperCAmelCase_ = tuple(state["cls"] ) UpperCAmelCase_ = False if state.get("add_prefix_space" , __a ) != add_prefix_space: UpperCAmelCase_ = add_prefix_space UpperCAmelCase_ = True if state.get("trim_offsets" , __a ) != trim_offsets: UpperCAmelCase_ = trim_offsets UpperCAmelCase_ = True if changes_to_apply: UpperCAmelCase_ = getattr(__a , state.pop("type" ) ) UpperCAmelCase_ = component_class(**__a ) setattr(self.backend_tokenizer , __a , __a ) @property def _lowercase (self : List[str] ): if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def _lowercase (self : str , __a : str ): UpperCAmelCase_ = AddedToken(__a , lstrip=__a , rstrip=__a ) if isinstance(__a , __a ) else value UpperCAmelCase_ = value def _lowercase (self : str , *__a : Dict , **__a : Optional[Any] ): UpperCAmelCase_ = kwargs.get("is_split_into_words" , __a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__a , **__a ) def _lowercase (self : Optional[int] , *__a : Optional[int] , **__a : int ): UpperCAmelCase_ = kwargs.get("is_split_into_words" , __a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._encode_plus(*__a , **__a ) def _lowercase (self : List[str] , __a : str , __a : Optional[str] = None ): UpperCAmelCase_ = self._tokenizer.model.save(__a , name=__a ) return tuple(__a ) def _lowercase (self : Tuple , __a : Any , __a : List[str]=None ): UpperCAmelCase_ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowercase (self : Tuple , __a : List[int] , __a : Optional[List[int]] = None ): UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
78
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=7, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=2, lowerCamelCase=99, lowerCamelCase=0, lowerCamelCase=32, lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=0.1, lowerCamelCase=0.1, lowerCamelCase=5_12, lowerCamelCase=2, lowerCamelCase=0.0_2, lowerCamelCase=2, lowerCamelCase=4, lowerCamelCase="last", lowerCamelCase=True, lowerCamelCase=None, lowerCamelCase=0, ) -> str: """simple docstring""" _lowercase : Union[str, Any] = parent _lowercase : Optional[Any] = batch_size _lowercase : List[str] = seq_length _lowercase : int = is_training _lowercase : List[str] = use_input_lengths _lowercase : int = use_token_type_ids _lowercase : Any = use_labels _lowercase : Union[str, Any] = gelu_activation _lowercase : List[str] = sinusoidal_embeddings _lowercase : str = causal _lowercase : Optional[int] = asm _lowercase : Union[str, Any] = n_langs _lowercase : List[Any] = vocab_size _lowercase : Any = n_special _lowercase : Any = hidden_size _lowercase : str = num_hidden_layers _lowercase : Union[str, Any] = num_attention_heads _lowercase : Tuple = hidden_dropout_prob _lowercase : Optional[int] = attention_probs_dropout_prob _lowercase : Union[str, Any] = max_position_embeddings _lowercase : List[str] = type_sequence_label_size _lowercase : Any = initializer_range _lowercase : int = num_labels _lowercase : Optional[int] = num_choices _lowercase : Optional[Any] = summary_type _lowercase : Optional[Any] = use_proj _lowercase : int = scope _lowercase : List[Any] = bos_token_id def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Tuple = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) _lowercase : List[str] = random_attention_mask([self.batch_size, self.seq_length]) _lowercase : int = None if self.use_input_lengths: _lowercase : Dict = ( ids_tensor([self.batch_size], vocab_size=2) + self.seq_length - 2 ) # small variation of seq_length _lowercase : Tuple = None if self.use_token_type_ids: _lowercase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length], self.n_langs) _lowercase : Tuple = None _lowercase : int = None _lowercase : int = None if self.use_labels: _lowercase : str = ids_tensor([self.batch_size], self.type_sequence_label_size) _lowercase : str = ids_tensor([self.batch_size, self.seq_length], self.num_labels) _lowercase : Dict = ids_tensor([self.batch_size], 2).float() _lowercase : Tuple = ids_tensor([self.batch_size], self.num_choices) _lowercase : Dict = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return XLMConfig( vocab_size=self.vocab_size, n_special=self.n_special, emb_dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, gelu_activation=self.gelu_activation, sinusoidal_embeddings=self.sinusoidal_embeddings, asm=self.asm, causal=self.causal, n_langs=self.n_langs, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, summary_type=self.summary_type, use_proj=self.use_proj, num_labels=self.num_labels, bos_token_id=self.bos_token_id, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Tuple: """simple docstring""" _lowercase : List[Any] = XLMModel(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = model(lowerCamelCase, lengths=lowerCamelCase, langs=lowerCamelCase) _lowercase : int = model(lowerCamelCase, langs=lowerCamelCase) _lowercase : Any = model(lowerCamelCase) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[Any]: """simple docstring""" _lowercase : Dict = XLMWithLMHeadModel(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> str: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnsweringSimple(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Dict = model(lowerCamelCase) _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) _lowercase : Any = outputs self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnswering(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = model(lowerCamelCase) _lowercase : List[Any] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, p_mask=lowerCamelCase, ) _lowercase : List[str] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, ) ((_lowercase) , ) : Optional[Any] = result_with_labels.to_tuple() _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) ((_lowercase) , ) : Any = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape, ()) self.parent.assertEqual(result.start_top_log_probs.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual(result.start_top_index.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual( result.end_top_log_probs.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual( result.end_top_index.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual(result.cls_logits.shape, (self.batch_size,)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> int: """simple docstring""" _lowercase : Optional[Any] = XLMForSequenceClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[int] = model(lowerCamelCase) _lowercase : Optional[int] = model(lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[str]: """simple docstring""" _lowercase : Any = self.num_labels _lowercase : str = XLMForTokenClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, attention_mask=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Dict: """simple docstring""" _lowercase : Optional[Any] = self.num_choices _lowercase : Optional[int] = XLMForMultipleChoice(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = input_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : int = token_type_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : Optional[Any] = input_mask.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : List[str] = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices)) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Dict = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[Any] = config_and_inputs _lowercase : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return config, inputs_dict @require_torch class _lowerCamelCase( _a, _a, _a, unittest.TestCase ): lowercase_ : Any = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) lowercase_ : Optional[int] = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable lowercase_ : Union[str, Any] = ( { """feature-extraction""": XLMModel, """fill-mask""": XLMWithLMHeadModel, """question-answering""": XLMForQuestionAnsweringSimple, """text-classification""": XLMForSequenceClassification, """text-generation""": XLMWithLMHeadModel, """token-classification""": XLMForTokenClassification, """zero-shot""": XLMForSequenceClassification, } if is_torch_available() else {} ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast') ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase=False) -> Optional[int]: """simple docstring""" _lowercase : Any = super()._prepare_for_class(lowerCamelCase, lowerCamelCase, return_labels=lowerCamelCase) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": _lowercase : Any = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) _lowercase : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) return inputs_dict def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Union[str, Any] = XLMModelTester(self) _lowercase : List[str] = ConfigTester(self, config_class=lowerCamelCase, emb_dim=37) def UpperCamelCase ( self) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*lowerCamelCase) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*lowerCamelCase) def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> int: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_attentions in attentions], [True] * len(lowerCamelCase)) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_attentions in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : Dict = min_length + idx + 1 _lowercase : int = min_length + idx + 1 _lowercase : Dict = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions], [expected_shape] * len(lowerCamelCase)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> List[Any]: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_hidden_states in hidden_states], [True] * len(lowerCamelCase), ) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_hidden_states in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : int = min_length + idx + 1 _lowercase : int = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states], [expected_shape] * len(lowerCamelCase), ) pass @slow def UpperCamelCase ( self) -> int: """simple docstring""" for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Dict = XLMModel.from_pretrained(lowerCamelCase) self.assertIsNotNone(lowerCamelCase) @require_torch class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMWithLMHeadModel.from_pretrained('xlm-mlm-en-2048') model.to(lowerCamelCase) _lowercase : Union[str, Any] = torch.tensor([[14, 4_47]], dtype=torch.long, device=lowerCamelCase) # the president _lowercase : Any = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference _lowercase : str = model.generate(lowerCamelCase, do_sample=lowerCamelCase) self.assertListEqual(output_ids[0].cpu().numpy().tolist(), lowerCamelCase)
89
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ : Any = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""PerceiverFeatureExtractor"""] SCREAMING_SNAKE_CASE__ : Any = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : int = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
79
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.17.0.dev0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/text-classification/requirements.txt") SCREAMING_SNAKE_CASE : Dict = logging.getLogger(__name__) @dataclass class _lowerCamelCase: lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""}, ) lowercase_ : int = field( default=10_24, metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of prediction examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the training data."""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the validation data."""} ) lowercase_ : Optional[str] = field(default=_a, metadata={"""help""": """A csv or a json file containing the test data."""} ) def UpperCamelCase ( self) -> Dict: """simple docstring""" if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError('Need either a GLUE task, a training/validation file or a dataset name.') else: _lowercase : int = self.train_file.split('.')[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." _lowercase : Tuple = self.validation_file.split('.')[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class _lowerCamelCase: lowercase_ : str = field( default=_a, metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""}, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""}, ) lowercase_ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) def UpperCamelCase_( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowercase , _lowercase , _lowercase : Tuple = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowercase , _lowercase , _lowercase : Union[str, Any] = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) _lowercase : Union[str, Any] = training_args.get_process_log_level() logger.setLevel(lowerCamelCase_ ) datasets.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _lowercase : Optional[int] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowercase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. _lowercase : Dict = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. _lowercase : Optional[Any] = {'train': data_args.train_file, 'validation': data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: _lowercase : Tuple = data_args.train_file.split('.' )[-1] _lowercase : int = data_args.test_file.split('.' )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." _lowercase : Any = data_args.test_file else: raise ValueError('Need either a GLUE task or a test file for `do_predict`.' ) for key in data_files.keys(): logger.info(F'''load a local file for {key}: {data_files[key]}''' ) if data_args.train_file.endswith('.csv' ): # Loading a dataset from local csv files _lowercase : str = load_dataset('csv' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files _lowercase : Optional[int] = load_dataset('json' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels _lowercase : Optional[Any] = raw_datasets['train'].features['label'].names _lowercase : Any = len(lowerCamelCase_ ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer _lowercase : str = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=lowerCamelCase_ , ) _lowercase : Tuple = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: _lowercase : int = 'max_length' else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _lowercase : str = False # Some models have set the order of the labels to use, so let's make sure we do use it. _lowercase : List[Any] = {'Refused': 0, 'Entailed': 1} _lowercase : Union[str, Any] = {0: 'Refused', 1: 'Entailed'} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) _lowercase : List[str] = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(lowerCamelCase_ ): # Tokenize the texts def _convert_table_text_to_pandas(lowerCamelCase_ ): _lowercase : int = [_table_row.split('#' ) for _table_row in _table_text.strip('\n' ).split('\n' )] _lowercase : Any = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd _lowercase : List[Any] = examples['statement'] _lowercase : Optional[Any] = list(map(_convert_table_text_to_pandas , examples['table_text'] ) ) _lowercase : Union[str, Any] = tokenizer(lowerCamelCase_ , lowerCamelCase_ , padding=lowerCamelCase_ , max_length=lowerCamelCase_ , truncation=lowerCamelCase_ ) _lowercase : Any = examples['label'] return result with training_args.main_process_first(desc='dataset map pre-processing' ): _lowercase : str = raw_datasets.map( lowerCamelCase_ , batched=lowerCamelCase_ , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on dataset' , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _lowercase : Any = raw_datasets['train'] if data_args.max_train_samples is not None: _lowercase : str = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _lowercase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _lowercase : List[Any] = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError('--do_predict requires a test dataset' ) _lowercase : Optional[int] = raw_datasets['test'] if data_args.max_predict_samples is not None: _lowercase : List[str] = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(lowerCamelCase_ ) ) , 3 ): logger.info(F'''Sample {index} of the training set: {train_dataset[index]}.''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCamelCase_ ): _lowercase : Dict = p.predictions[0] if isinstance(p.predictions , lowerCamelCase_ ) else p.predictions _lowercase : Tuple = np.argmax(lowerCamelCase_ , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _lowercase : Any = default_data_collator elif training_args.fpaa: _lowercase : str = DataCollatorWithPadding(lowerCamelCase_ , pad_to_multiple_of=8 ) else: _lowercase : Optional[Any] = None # Initialize our Trainer _lowercase : List[str] = Trainer( model=lowerCamelCase_ , args=lowerCamelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=lowerCamelCase_ , tokenizer=lowerCamelCase_ , data_collator=lowerCamelCase_ , ) # Training if training_args.do_train: _lowercase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _lowercase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowercase : Optional[Any] = last_checkpoint _lowercase : Optional[Any] = trainer.train(resume_from_checkpoint=lowerCamelCase_ ) _lowercase : List[Any] = train_result.metrics _lowercase : Dict = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCamelCase_ ) ) _lowercase : int = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics('train' , lowerCamelCase_ ) trainer.save_metrics('train' , lowerCamelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _lowercase : Tuple = trainer.evaluate(eval_dataset=lowerCamelCase_ ) _lowercase : Any = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCamelCase_ ) _lowercase : Optional[int] = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.log_metrics('eval' , lowerCamelCase_ ) trainer.save_metrics('eval' , lowerCamelCase_ ) if training_args.do_predict: logger.info('*** Predict ***' ) # Removing the `label` columns because it contains -1 and Trainer won't like that. _lowercase : Any = predict_dataset.remove_columns('label' ) _lowercase : Optional[Any] = trainer.predict(lowerCamelCase_ , metric_key_prefix='predict' ).predictions _lowercase : Union[str, Any] = np.argmax(lowerCamelCase_ , axis=1 ) _lowercase : Dict = os.path.join(training_args.output_dir , 'predict_results_tabfact.txt' ) if trainer.is_world_process_zero(): with open(lowerCamelCase_ , 'w' ) as writer: logger.info('***** Predict Results *****' ) writer.write('index\tprediction\n' ) for index, item in enumerate(lowerCamelCase_ ): _lowercase : List[str] = label_list[item] writer.write(F'''{index}\t{item}\n''' ) _lowercase : str = {'finetuned_from': model_args.model_name_or_path, 'tasks': 'text-classification'} if training_args.push_to_hub: trainer.push_to_hub(**lowerCamelCase_ ) else: trainer.create_model_card(**lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
89
0
from math import factorial def snake_case ( lowerCamelCase , lowerCamelCase ): '''simple docstring''' if n < k or k < 0: raise ValueError("""Please enter positive integers for n and k where n >= k""" ) return factorial(lowerCamelCase ) // (factorial(lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( """The number of five-card hands possible from a standard""", F'''fifty-two card deck is: {combinations(52, 5)}\n''', ) print( """If a class of 40 students must be arranged into groups of""", F'''4 for group projects, there are {combinations(40, 4)} ways''', """to arrange them.\n""", ) print( """If 10 teams are competing in a Formula One race, there""", F'''are {combinations(10, 3)} ways that first, second and''', """third place can be awarded.""", )
80
from maths.prime_factors import prime_factors def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): _lowercase : str = F'''Input value of [number={number}] must be an integer''' raise TypeError(lowerCamelCase_ ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(lowerCamelCase_ ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
89
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer _snake_case : List[str] = logging.get_logger(__name__) _snake_case : Dict = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _snake_case : Any = { "vocab_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json" ), "distilbert-base-german-cased": ( "https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json" ), "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json" ), }, } _snake_case : str = { "distilbert-base-uncased": 512, "distilbert-base-uncased-distilled-squad": 512, "distilbert-base-cased": 512, "distilbert-base-cased-distilled-squad": 512, "distilbert-base-german-cased": 512, "distilbert-base-multilingual-cased": 512, } _snake_case : Dict = { "distilbert-base-uncased": {"do_lower_case": True}, "distilbert-base-uncased-distilled-squad": {"do_lower_case": True}, "distilbert-base-cased": {"do_lower_case": False}, "distilbert-base-cased-distilled-squad": {"do_lower_case": False}, "distilbert-base-german-cased": {"do_lower_case": False}, "distilbert-base-multilingual-cased": {"do_lower_case": False}, } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Dict = VOCAB_FILES_NAMES __UpperCAmelCase : int = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Optional[Any] = PRETRAINED_INIT_CONFIGURATION __UpperCAmelCase : int = ["input_ids", "attention_mask"] __UpperCAmelCase : Tuple = DistilBertTokenizer def __init__( self : int , lowerCamelCase : Any=None , lowerCamelCase : int=None , lowerCamelCase : Tuple=True , lowerCamelCase : int="[UNK]" , lowerCamelCase : int="[SEP]" , lowerCamelCase : Optional[Any]="[PAD]" , lowerCamelCase : List[str]="[CLS]" , lowerCamelCase : List[Any]="[MASK]" , lowerCamelCase : Dict=True , lowerCamelCase : str=None , **lowerCamelCase : int , ) -> List[Any]: super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , do_lower_case=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , pad_token=lowerCamelCase , cls_token=lowerCamelCase , mask_token=lowerCamelCase , tokenize_chinese_chars=lowerCamelCase , strip_accents=lowerCamelCase , **lowerCamelCase , ) __snake_case : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCamelCase ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCamelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCamelCase ) != tokenize_chinese_chars ): __snake_case : List[Any] = getattr(lowerCamelCase , normalizer_state.pop("type" ) ) __snake_case : Union[str, Any] = do_lower_case __snake_case : Union[str, Any] = strip_accents __snake_case : List[str] = tokenize_chinese_chars __snake_case : Optional[Any] = normalizer_class(**lowerCamelCase ) __snake_case : List[Any] = do_lower_case def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any=None ) -> Any: __snake_case : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __snake_case ( self : List[Any] , lowerCamelCase : List[int] , lowerCamelCase : Optional[List[int]] = None ) -> List[int]: __snake_case : Union[str, Any] = [self.sep_token_id] __snake_case : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __snake_case ( self : str , lowerCamelCase : str , lowerCamelCase : Optional[str] = None ) -> Tuple[str]: __snake_case : str = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase )
81
from __future__ import annotations from typing import Any class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = 0) -> None: """simple docstring""" _lowercase , _lowercase : str = row, column _lowercase : Any = [[default_value for c in range(lowerCamelCase)] for r in range(lowerCamelCase)] def __str__( self) -> str: """simple docstring""" _lowercase : Tuple = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _lowercase : str = 0 for row_vector in self.array: for obj in row_vector: _lowercase : Optional[int] = max(lowerCamelCase, len(str(lowerCamelCase))) _lowercase : List[str] = F'''%{max_element_length}s''' # Make string and return def single_line(lowerCamelCase) -> str: nonlocal string_format_identifier _lowercase : Union[str, Any] = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector) line += "]" return line s += "\n".join(single_line(lowerCamelCase) for row_vector in self.array) return s def __repr__( self) -> str: """simple docstring""" return str(self) def UpperCamelCase ( self, lowerCamelCase) -> bool: """simple docstring""" if not (isinstance(lowerCamelCase, (list, tuple)) and len(lowerCamelCase) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self, lowerCamelCase) -> Any: """simple docstring""" assert self.validate_indicies(lowerCamelCase) return self.array[loc[0]][loc[1]] def __setitem__( self, lowerCamelCase, lowerCamelCase) -> None: """simple docstring""" assert self.validate_indicies(lowerCamelCase) _lowercase : Optional[Any] = value def __add__( self, lowerCamelCase) -> Matrix: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) assert self.row == another.row and self.column == another.column # Add _lowercase : Any = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : int = self[r, c] + another[r, c] return result def __neg__( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : List[str] = -self[r, c] return result def __sub__( self, lowerCamelCase) -> Matrix: """simple docstring""" return self + (-another) def __mul__( self, lowerCamelCase) -> Matrix: """simple docstring""" if isinstance(lowerCamelCase, (int, float)): # Scalar multiplication _lowercase : Dict = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] * another return result elif isinstance(lowerCamelCase, lowerCamelCase): # Matrix multiplication assert self.column == another.row _lowercase : str = Matrix(self.row, another.column) for r in range(self.row): for c in range(another.column): for i in range(self.column): result[r, c] += self[r, i] * another[i, c] return result else: _lowercase : Tuple = F'''Unsupported type given for another ({type(lowerCamelCase)})''' raise TypeError(lowerCamelCase) def UpperCamelCase ( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.column, self.row) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] return result def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> Any: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) and isinstance(lowerCamelCase, lowerCamelCase) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _lowercase : Dict = v.transpose() _lowercase : Any = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def UpperCamelCase_( ) -> None: # a^(-1) _lowercase : Optional[int] = Matrix(3 , 3 , 0 ) for i in range(3 ): _lowercase : int = 1 print(F'''a^(-1) is {ainv}''' ) # u, v _lowercase : Dict = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : Dict = 1, 2, -3 _lowercase : List[Any] = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : int = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCamelCase_ , lowerCamelCase_ )}''' ) def UpperCamelCase_( ) -> None: import doctest doctest.testmod() testa()
89
0
"""simple docstring""" def a__ ( lowerCAmelCase__ ): if number > 0: raise ValueError("input must be a negative integer" ) UpperCAmelCase_ = len(bin(lowerCAmelCase__ )[3:] ) UpperCAmelCase_ = bin(abs(lowerCAmelCase__ ) - (1 << binary_number_length) )[3:] UpperCAmelCase_ = ( ( "1" + "0" * (binary_number_length - len(lowerCAmelCase__ )) + twos_complement_number ) if number < 0 else "0" ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
82
import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = int(lowerCamelCase_ ) _lowercase , _lowercase , _lowercase : Optional[Any] = t // 3600, (t // 60) % 60, t % 60 return F'''{h}:{m:02d}:{s:02d}''' if h != 0 else F'''{m:02d}:{s:02d}''' def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=300 ) -> Dict: # docstyle-ignore return F''' <div> {prefix} <progress value=\'{value}\' max=\'{total}\' style=\'width:{width}px; height:20px; vertical-align: middle;\'></progress> {label} </div> ''' def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : int = '<table border="1" class="dataframe">\n' html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += F''' <th>{i}</th>\n''' html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: _lowercase : Any = F'''{elt:.6f}''' if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else str(lowerCamelCase_ ) html_code += F''' <td>{elt}</td>\n''' html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class _lowerCamelCase: lowercase_ : str = 5 lowercase_ : str = 0.2 def __init__( self, lowerCamelCase, lowerCamelCase = None, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = 3_00, ) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = total _lowercase : Optional[int] = '' if prefix is None else prefix _lowercase : Tuple = leave _lowercase : str = parent _lowercase : str = width _lowercase : List[Any] = None _lowercase : List[str] = None _lowercase : Tuple = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, lowerCamelCase = None) -> Dict: """simple docstring""" _lowercase : Any = value if comment is not None: _lowercase : Union[str, Any] = comment if self.last_value is None: _lowercase : Dict = time.time() _lowercase : Tuple = value _lowercase : str = None _lowercase : Optional[int] = self.warmup _lowercase : Optional[Any] = 1 self.update_bar(lowerCamelCase) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for, self.total): if self.first_calls > 0: self.first_calls -= 1 _lowercase : List[str] = time.time() _lowercase : Tuple = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: _lowercase : Dict = self.elapsed_time / (value - self.start_value) else: _lowercase : int = None if value >= self.total: _lowercase : Dict = self.total _lowercase : List[str] = None if not self.leave: self.close() elif self.average_time_per_item is not None: _lowercase : Optional[int] = self.average_time_per_item * (self.total - value) self.update_bar(lowerCamelCase) _lowercase : int = value _lowercase : Tuple = current_time if self.average_time_per_item is None: _lowercase : str = 1 else: _lowercase : int = max(int(self.update_every / self.average_time_per_item), 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None) -> Optional[Any]: """simple docstring""" _lowercase : List[Any] = ' ' * (len(str(self.total)) - len(str(lowerCamelCase))) + str(lowerCamelCase) if self.elapsed_time is None: _lowercase : int = F'''[{spaced_value}/{self.total} : < :''' elif self.predicted_remaining is None: _lowercase : Union[str, Any] = F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)}''' else: _lowercase : Union[str, Any] = ( F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)} <''' F''' {format_time(self.predicted_remaining)}''' ) self.label += F''', {1/self.average_time_per_item:.2f} it/s''' self.label += "]" if self.comment is None or len(self.comment) == 0 else F''', {self.comment}]''' self.display() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: _lowercase : Optional[Any] = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" if self.parent is None and self.output is not None: self.output.update(disp.HTML('')) class _lowerCamelCase( _a ): def __init__( self, lowerCamelCase, lowerCamelCase=None) -> int: """simple docstring""" super().__init__(lowerCamelCase) _lowercase : Optional[Any] = None if column_names is None else [column_names] _lowercase : Any = None def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: _lowercase : Dict = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self, lowerCamelCase) -> Dict: """simple docstring""" if self.inner_table is None: _lowercase : Dict = [list(values.keys()), list(values.values())] else: _lowercase : Tuple = self.inner_table[0] if len(self.inner_table) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(lowerCamelCase) _lowercase : str = columns self.inner_table.append([values[c] for c in columns]) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None, lowerCamelCase=3_00) -> Union[str, Any]: """simple docstring""" _lowercase : List[str] = NotebookProgressBar(lowerCamelCase, prefix=lowerCamelCase, parent=self, width=lowerCamelCase) return self.child_bar def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Optional[Any] = None self.display() class _lowerCamelCase( _a ): def __init__( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = None _lowercase : Dict = None _lowercase : Dict = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Dict: """simple docstring""" _lowercase : Dict = 'Epoch' if args.evaluation_strategy == IntervalStrategy.EPOCH else 'Step' _lowercase : Dict = 0 _lowercase : Tuple = 0 _lowercase : int = [self.first_column] + ['Training Loss'] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append('Validation Loss') _lowercase : Union[str, Any] = NotebookTrainingTracker(state.max_steps, lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" _lowercase : Any = int(state.epoch) if int(state.epoch) == state.epoch else F'''{state.epoch:.2f}''' self.training_tracker.update( state.global_step + 1, comment=F'''Epoch {epoch}/{state.num_train_epochs}''', force_update=self._force_next_update, ) _lowercase : str = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Any: """simple docstring""" if not has_length(lowerCamelCase): return if self.prediction_bar is None: if self.training_tracker is not None: _lowercase : Optional[int] = self.training_tracker.add_child(len(lowerCamelCase)) else: _lowercase : Optional[int] = NotebookProgressBar(len(lowerCamelCase)) self.prediction_bar.update(1) else: self.prediction_bar.update(self.prediction_bar.value + 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Optional[int]: """simple docstring""" if self.prediction_bar is not None: self.prediction_bar.close() _lowercase : Any = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[Any]: """simple docstring""" if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: _lowercase : Dict = {'Training Loss': logs['loss']} # First column is necessarily Step sine we're not in epoch eval strategy _lowercase : List[Any] = state.global_step self.training_tracker.write_line(lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[str]: """simple docstring""" if self.training_tracker is not None: _lowercase : Tuple = {'Training Loss': 'No log', 'Validation Loss': 'No log'} for log in reversed(state.log_history): if "loss" in log: _lowercase : int = log['loss'] break if self.first_column == "Epoch": _lowercase : Union[str, Any] = int(state.epoch) else: _lowercase : Optional[Any] = state.global_step _lowercase : str = 'eval' for k in metrics: if k.endswith('_loss'): _lowercase : str = re.sub(R'\_loss$', '', lowerCamelCase) _lowercase : Tuple = metrics.pop('total_flos', lowerCamelCase) _lowercase : List[str] = metrics.pop('epoch', lowerCamelCase) _lowercase : List[Any] = metrics.pop(F'''{metric_key_prefix}_runtime''', lowerCamelCase) _lowercase : Dict = metrics.pop(F'''{metric_key_prefix}_samples_per_second''', lowerCamelCase) _lowercase : Tuple = metrics.pop(F'''{metric_key_prefix}_steps_per_second''', lowerCamelCase) _lowercase : List[str] = metrics.pop(F'''{metric_key_prefix}_jit_compilation_time''', lowerCamelCase) for k, v in metrics.items(): if k == F'''{metric_key_prefix}_loss''': _lowercase : Union[str, Any] = v else: _lowercase : Optional[Any] = k.split('_') _lowercase : Optional[int] = ' '.join([part.capitalize() for part in splits[1:]]) _lowercase : Tuple = v self.training_tracker.write_line(lowerCamelCase) self.training_tracker.remove_child() _lowercase : str = None # Evaluation takes a long time so we should force the next update. _lowercase : Optional[Any] = True def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" self.training_tracker.update( state.global_step, comment=F'''Epoch {int(state.epoch)}/{state.num_train_epochs}''', force_update=lowerCamelCase) _lowercase : Any = None
89
0
"""simple docstring""" from __future__ import annotations lowerCAmelCase__ = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] lowerCAmelCase__ = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def snake_case_ ( A_ : list[float] ): '''simple docstring''' _lowerCamelCase : Tuple = [] _lowerCamelCase : List[str] = len(A_ ) for i in range(A_ ): _lowerCamelCase : float = -1 for j in range(i + 1, A_ ): if arr[i] < arr[j]: _lowerCamelCase : int = arr[j] break result.append(A_ ) return result def snake_case_ ( A_ : list[float] ): '''simple docstring''' _lowerCamelCase : Optional[Any] = [] for i, outer in enumerate(A_ ): _lowerCamelCase : float = -1 for inner in arr[i + 1 :]: if outer < inner: _lowerCamelCase : List[str] = inner break result.append(A_ ) return result def snake_case_ ( A_ : list[float] ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = len(A_ ) _lowerCamelCase : list[float] = [] _lowerCamelCase : list[float] = [-1] * arr_size for index in reversed(range(A_ ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _lowerCamelCase : Any = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) lowerCAmelCase__ = ( '''from __main__ import arr, next_greatest_element_slow, ''' '''next_greatest_element_fast, next_greatest_element''' ) print( '''next_greatest_element_slow():''', timeit('''next_greatest_element_slow(arr)''', setup=setup), ) print( '''next_greatest_element_fast():''', timeit('''next_greatest_element_fast(arr)''', setup=setup), ) print( ''' next_greatest_element():''', timeit('''next_greatest_element(arr)''', setup=setup), )
83
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 18, 2] _lowercase : Tuple = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Any = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Dict = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: _lowercase : Any = [3, 3, 3, 3] _lowercase : Any = [5, 5, 5, 5] elif "fl4" in model_name: _lowercase : Dict = [4, 4, 4, 4] _lowercase : Tuple = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: _lowercase : str = [3, 3, 3, 3] if "lrf" in model_name: _lowercase : Optional[int] = [3, 3, 3, 3] else: _lowercase : Dict = [2, 2, 2, 2] if "tiny" in model_name: _lowercase : List[str] = 96 elif "small" in model_name: _lowercase : Dict = 96 elif "base" in model_name: _lowercase : Optional[int] = 128 elif "large" in model_name: _lowercase : List[Any] = 192 elif "xlarge" in model_name: _lowercase : Optional[Any] = 256 elif "huge" in model_name: _lowercase : Dict = 352 # set label information _lowercase : int = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: _lowercase : str = 'imagenet-22k-id2label.json' else: _lowercase : Tuple = 'imagenet-1k-id2label.json' _lowercase : Union[str, Any] = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : Any = {v: k for k, v in idalabel.items()} _lowercase : Optional[Any] = FocalNetConfig( embed_dim=lowerCamelCase_ , depths=lowerCamelCase_ , focal_levels=lowerCamelCase_ , focal_windows=lowerCamelCase_ , use_conv_embed=lowerCamelCase_ , idalabel=lowerCamelCase_ , labelaid=lowerCamelCase_ , use_post_layernorm=lowerCamelCase_ , use_layerscale=lowerCamelCase_ , ) return config def UpperCamelCase_( lowerCamelCase_ ) -> Any: if "patch_embed.proj" in name: _lowercase : Optional[Any] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: _lowercase : str = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: _lowercase : Any = 'encoder.' + name if "encoder.layers" in name: _lowercase : int = name.replace('encoder.layers' , 'encoder.stages' ) if "downsample.proj" in name: _lowercase : Tuple = name.replace('downsample.proj' , 'downsample.projection' ) if "blocks" in name: _lowercase : str = name.replace('blocks' , 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: _lowercase : List[str] = name.replace('modulation.f' , 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: _lowercase : int = name.replace('modulation.h' , 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: _lowercase : Any = name.replace('modulation.proj' , 'modulation.projection_out' ) if name == "norm.weight": _lowercase : Any = 'layernorm.weight' if name == "norm.bias": _lowercase : Tuple = 'layernorm.bias' if "head" in name: _lowercase : Optional[int] = name.replace('head' , 'classifier' ) else: _lowercase : Optional[int] = 'focalnet.' + name return name def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False ) -> str: # fmt: off _lowercase : Dict = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on _lowercase : Dict = model_name_to_url[model_name] print('Checkpoint URL: ' , lowerCamelCase_ ) _lowercase : List[str] = torch.hub.load_state_dict_from_url(lowerCamelCase_ , map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[int] = val _lowercase : Union[str, Any] = get_focalnet_config(lowerCamelCase_ ) _lowercase : Optional[Any] = FocalNetForImageClassification(lowerCamelCase_ ) model.eval() # load state dict model.load_state_dict(lowerCamelCase_ ) # verify conversion _lowercase : Optional[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : Any = BitImageProcessor( do_resize=lowerCamelCase_ , size={'shortest_edge': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=lowerCamelCase_ , crop_size=224 , do_normalize=lowerCamelCase_ , image_mean=lowerCamelCase_ , image_std=lowerCamelCase_ , ) _lowercase : List[str] = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) _lowercase : List[Any] = processor(images=lowerCamelCase_ , return_tensors='pt' ) _lowercase : str = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ), ] ) _lowercase : List[str] = image_transforms(lowerCamelCase_ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , lowerCamelCase_ , atol=1e-4 ) _lowercase : Dict = model(**lowerCamelCase_ ) _lowercase : int = outputs.logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) print('First values of logits:' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": _lowercase : Optional[Any] = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": _lowercase : int = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": _lowercase : str = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": _lowercase : Any = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": _lowercase : List[Any] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": _lowercase : int = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3] , lowerCamelCase_ , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=2 , snake_case=24 , snake_case=16 , snake_case=True , snake_case=True , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=10 , snake_case=0.02 , snake_case=None , snake_case=2 , snake_case=2 , ): lowercase = parent lowercase = batch_size lowercase = patch_size lowercase = max_length lowercase = num_mel_bins lowercase = is_training lowercase = use_labels lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = type_sequence_label_size lowercase = initializer_range lowercase = scope lowercase = frequency_stride lowercase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowercase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 lowercase = (self.max_length - self.patch_size) // self.time_stride + 1 lowercase = frequency_out_dimension * time_out_dimension lowercase = num_patches + 2 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = self.get_config() return config, input_values, labels def SCREAMING_SNAKE_CASE__ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): lowercase = ASTModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'input_values': input_values} return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : str = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) _UpperCamelCase : Optional[int] = ( {"""audio-classification""": ASTForAudioClassification, """feature-extraction""": ASTModel} if is_torch_available() else {} ) _UpperCamelCase : Union[str, Any] = False _UpperCamelCase : List[Any] = False _UpperCamelCase : Any = False _UpperCamelCase : Optional[int] = False def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ASTModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , has_text_modality=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = model_class(snake_case ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case , nn.Linear ) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = model_class(snake_case ) lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase = [*signature.parameters.keys()] lowercase = ['input_values'] self.assertListEqual(arg_names[:1] , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = ASTModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def UpperCAmelCase_ ( ): lowercase = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' ) lowercase , lowercase = torchaudio.load(__SCREAMING_SNAKE_CASE ) return audio, sampling_rate @require_torch @require_torchaudio class A_ ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE__ ( self ): return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.default_feature_extractor lowercase = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(snake_case ) lowercase = self.default_feature_extractor lowercase , lowercase = prepare_audio() lowercase = audio.squeeze().numpy() lowercase = feature_extractor(snake_case , sampling_rate=snake_case , return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): lowercase = model(**snake_case ) # verify the logits lowercase = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , snake_case ) lowercase = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ).to(snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case , atol=1E-4 ) )
84
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Any = { "ut/deta": "https://huggingface.co/ut/deta/resolve/main/config.json", } class _lowerCamelCase( _a ): lowercase_ : Any = """deta""" lowercase_ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self, lowerCamelCase=None, lowerCamelCase=9_00, lowerCamelCase=20_48, lowerCamelCase=6, lowerCamelCase=20_48, lowerCamelCase=8, lowerCamelCase=6, lowerCamelCase=10_24, lowerCamelCase=8, lowerCamelCase=0.0, lowerCamelCase=True, lowerCamelCase="relu", lowerCamelCase=2_56, lowerCamelCase=0.1, lowerCamelCase=0.0, lowerCamelCase=0.0, lowerCamelCase=0.0_2, lowerCamelCase=1.0, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase="sine", lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=4, lowerCamelCase=True, lowerCamelCase=3_00, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=1, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=0.1, lowerCamelCase=0.2_5, **lowerCamelCase, ) -> Any: """simple docstring""" if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.') _lowercase : List[Any] = CONFIG_MAPPING['resnet'](out_features=['stage2', 'stage3', 'stage4']) else: if isinstance(lowerCamelCase, lowerCamelCase): _lowercase : Dict = backbone_config.pop('model_type') _lowercase : int = CONFIG_MAPPING[backbone_model_type] _lowercase : Union[str, Any] = config_class.from_dict(lowerCamelCase) _lowercase : Union[str, Any] = backbone_config _lowercase : Any = num_queries _lowercase : Union[str, Any] = max_position_embeddings _lowercase : Union[str, Any] = d_model _lowercase : Optional[int] = encoder_ffn_dim _lowercase : Optional[int] = encoder_layers _lowercase : Optional[Any] = encoder_attention_heads _lowercase : Optional[Any] = decoder_ffn_dim _lowercase : Dict = decoder_layers _lowercase : Tuple = decoder_attention_heads _lowercase : Union[str, Any] = dropout _lowercase : Optional[Any] = attention_dropout _lowercase : int = activation_dropout _lowercase : Tuple = activation_function _lowercase : List[Any] = init_std _lowercase : Union[str, Any] = init_xavier_std _lowercase : int = encoder_layerdrop _lowercase : Optional[int] = auxiliary_loss _lowercase : Dict = position_embedding_type # deformable attributes _lowercase : Any = num_feature_levels _lowercase : str = encoder_n_points _lowercase : Any = decoder_n_points _lowercase : List[str] = two_stage _lowercase : Dict = two_stage_num_proposals _lowercase : Any = with_box_refine _lowercase : List[Any] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.') # Hungarian matcher _lowercase : List[Any] = class_cost _lowercase : Optional[int] = bbox_cost _lowercase : str = giou_cost # Loss coefficients _lowercase : Optional[int] = mask_loss_coefficient _lowercase : int = dice_loss_coefficient _lowercase : List[Any] = bbox_loss_coefficient _lowercase : Optional[Any] = giou_loss_coefficient _lowercase : str = eos_coefficient _lowercase : int = focal_alpha super().__init__(is_encoder_decoder=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.d_model def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : int = copy.deepcopy(self.__dict__) _lowercase : Optional[int] = self.backbone_config.to_dict() _lowercase : Optional[Any] = self.__class__.model_type return output
89
0
from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging SCREAMING_SNAKE_CASE__ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : List[Any] = { "deepmind/language-perceiver": "https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json", # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class snake_case ( UpperCamelCase_ ): lowercase_ = 'perceiver' def __init__( self : List[Any] , a_ : List[Any]=256 , a_ : List[Any]=1280 , a_ : str=768 , a_ : List[Any]=1 , a_ : Tuple=26 , a_ : Optional[Any]=8 , a_ : str=8 , a_ : int=None , a_ : Dict=None , a_ : Dict="kv" , a_ : List[Any]=1 , a_ : Dict=1 , a_ : Dict="gelu" , a_ : Dict=0.1 , a_ : Optional[Any]=0.02 , a_ : Dict=1e-1_2 , a_ : Union[str, Any]=True , a_ : Optional[int]=262 , a_ : str=2048 , a_ : List[str]=56 , a_ : Dict=[368, 496] , a_ : int=16 , a_ : str=1920 , a_ : List[str]=16 , a_ : List[str]=[1, 16, 224, 224] , **a_ : Union[str, Any] , )-> Tuple: """simple docstring""" super().__init__(**a_ ) SCREAMING_SNAKE_CASE__ : Optional[int] = num_latents SCREAMING_SNAKE_CASE__ : Dict = d_latents SCREAMING_SNAKE_CASE__ : Optional[Any] = d_model SCREAMING_SNAKE_CASE__ : str = num_blocks SCREAMING_SNAKE_CASE__ : Tuple = num_self_attends_per_block SCREAMING_SNAKE_CASE__ : List[str] = num_self_attention_heads SCREAMING_SNAKE_CASE__ : Tuple = num_cross_attention_heads SCREAMING_SNAKE_CASE__ : Any = qk_channels SCREAMING_SNAKE_CASE__ : Tuple = v_channels SCREAMING_SNAKE_CASE__ : Any = cross_attention_shape_for_attention SCREAMING_SNAKE_CASE__ : List[str] = self_attention_widening_factor SCREAMING_SNAKE_CASE__ : Dict = cross_attention_widening_factor SCREAMING_SNAKE_CASE__ : Optional[Any] = hidden_act SCREAMING_SNAKE_CASE__ : List[str] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : Optional[Any] = initializer_range SCREAMING_SNAKE_CASE__ : Union[str, Any] = layer_norm_eps SCREAMING_SNAKE_CASE__ : Tuple = use_query_residual # masked language modeling attributes SCREAMING_SNAKE_CASE__ : Optional[Any] = vocab_size SCREAMING_SNAKE_CASE__ : str = max_position_embeddings # image classification attributes SCREAMING_SNAKE_CASE__ : List[Any] = image_size # flow attributes SCREAMING_SNAKE_CASE__ : Dict = train_size # multimodal autoencoding attributes SCREAMING_SNAKE_CASE__ : Tuple = num_frames SCREAMING_SNAKE_CASE__ : int = audio_samples_per_frame SCREAMING_SNAKE_CASE__ : Union[str, Any] = samples_per_patch SCREAMING_SNAKE_CASE__ : Optional[int] = output_shape class snake_case ( UpperCamelCase_ ): @property def __lowercase( self : str )-> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": SCREAMING_SNAKE_CASE__ : Optional[int] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: SCREAMING_SNAKE_CASE__ : List[str] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('inputs', dynamic_axis), ('attention_mask', dynamic_axis), ] ) @property def __lowercase( self : Tuple )-> float: """simple docstring""" return 1e-4 def __lowercase( self : List[str] , a_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , a_ : int = -1 , a_ : int = -1 , a_ : int = -1 , a_ : bool = False , a_ : Optional[TensorType] = None , a_ : int = 3 , a_ : int = 40 , a_ : int = 40 , )-> Mapping[str, Any]: """simple docstring""" # copied from `transformers.onnx.config.OnnxConfig` and slightly altered/simplified if isinstance(a_ , a_ ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX SCREAMING_SNAKE_CASE__ : List[Any] = compute_effective_axis_dimension( a_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX SCREAMING_SNAKE_CASE__ : Any = preprocessor.num_special_tokens_to_add(a_ ) SCREAMING_SNAKE_CASE__ : str = compute_effective_axis_dimension( a_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=a_ ) # Generate dummy inputs according to compute batch and sequence SCREAMING_SNAKE_CASE__ : Union[str, Any] = [' '.join(['a'] ) * seq_length] * batch_size SCREAMING_SNAKE_CASE__ : Tuple = dict(preprocessor(a_ , return_tensors=a_ ) ) SCREAMING_SNAKE_CASE__ : Dict = inputs.pop('input_ids' ) return inputs elif isinstance(a_ , a_ ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX SCREAMING_SNAKE_CASE__ : Dict = compute_effective_axis_dimension(a_ , fixed_dimension=OnnxConfig.default_fixed_batch ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = self._generate_dummy_images(a_ , a_ , a_ , a_ ) SCREAMING_SNAKE_CASE__ : Tuple = dict(preprocessor(images=a_ , return_tensors=a_ ) ) SCREAMING_SNAKE_CASE__ : str = inputs.pop('pixel_values' ) return inputs else: raise ValueError( 'Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor.' )
85
from __future__ import annotations import numpy as np def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: return np.maximum(0 , lowerCamelCase_ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
89
0
def __snake_case ( __UpperCamelCase : str ): """simple docstring""" return credit_card_number.startswith(("34", "35", "37", "4", "5", "6") ) def __snake_case ( __UpperCamelCase : str ): """simple docstring""" A_ = credit_card_number A_ = 0 A_ = len(__UpperCamelCase ) - 2 for i in range(__UpperCamelCase ,-1 ,-2 ): # double the value of every second digit A_ = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 A_ = cc_number[:i] + str(__UpperCamelCase ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(__UpperCamelCase ) - 1 ,-1 ,-2 ): total += int(cc_number[i] ) return total % 10 == 0 def __snake_case ( __UpperCamelCase : str ): """simple docstring""" A_ = f'''{credit_card_number} is an invalid credit card number because''' if not credit_card_number.isdigit(): print(f'''{error_message} it has nonnumerical characters.''' ) return False if not 13 <= len(__UpperCamelCase ) <= 16: print(f'''{error_message} of its length.''' ) return False if not validate_initial_digits(__UpperCamelCase ): print(f'''{error_message} of its first two digits.''' ) return False if not luhn_validation(__UpperCamelCase ): print(f'''{error_message} it fails the Luhn check.''' ) return False print(f'''{credit_card_number} is a valid credit card number.''' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
86
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: # Initialise PyTorch model _lowercase : Optional[int] = TaConfig.from_json_file(lowerCamelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase : Union[str, Any] = TaForConditionalGeneration(lowerCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
89
0
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType _lowerCamelCase : Optional[List[str]] = None _lowerCamelCase : int = """<""" if sys.byteorder == """little""" else """>""" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image _lowerCamelCase : Union[str, Any] = [ np.dtype("""|b1"""), np.dtype("""|u1"""), np.dtype("""<u2"""), np.dtype(""">u2"""), np.dtype("""<i2"""), np.dtype(""">i2"""), np.dtype("""<u4"""), np.dtype(""">u4"""), np.dtype("""<i4"""), np.dtype(""">i4"""), np.dtype("""<f4"""), np.dtype(""">f4"""), np.dtype("""<f8"""), np.dtype(""">f8"""), ] @dataclass class UpperCamelCase_ : '''simple docstring''' UpperCAmelCase__ = True UpperCAmelCase__ = None # Automatically constructed UpperCAmelCase__ = "PIL.Image.Image" UpperCAmelCase__ = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) UpperCAmelCase__ = field(default='''Image''' , init=UpperCAmelCase__ , repr=UpperCAmelCase__ ) def __call__( self : List[str]) ->List[str]: '''simple docstring''' return self.pa_type def SCREAMING_SNAKE_CASE ( self : Optional[Any] , UpperCAmelCase__ : Union[str, bytes, dict, np.ndarray, "PIL.Image.Image"]) ->dict: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''') if isinstance(UpperCAmelCase__ , UpperCAmelCase__): A__ = np.array(UpperCAmelCase__) if isinstance(UpperCAmelCase__ , UpperCAmelCase__): return {"path": value, "bytes": None} elif isinstance(UpperCAmelCase__ , UpperCAmelCase__): return {"path": None, "bytes": value} elif isinstance(UpperCAmelCase__ , np.ndarray): # convert the image array to PNG/TIFF bytes return encode_np_array(UpperCAmelCase__) elif isinstance(UpperCAmelCase__ , PIL.Image.Image): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(UpperCAmelCase__) elif value.get('''path''') is not None and os.path.isfile(value['''path''']): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get('''path''')} elif value.get('''bytes''') is not None or value.get('''path''') is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get('''bytes'''), "path": value.get('''path''')} else: raise ValueError( f"""An image sample should have one of 'path' or 'bytes' but they are missing or None in {value}.""") def SCREAMING_SNAKE_CASE ( self : Optional[int] , UpperCAmelCase__ : dict , UpperCAmelCase__ : str=None) ->"PIL.Image.Image": '''simple docstring''' if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Image(decode=True) instead.''') if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support decoding images, please install \'Pillow\'.''') if token_per_repo_id is None: A__ = {} A__ , A__ = value['''path'''], value['''bytes'''] if bytes_ is None: if path is None: raise ValueError(f"""An image should have one of 'path' or 'bytes' but both are None in {value}.""") else: if is_local_path(UpperCAmelCase__): A__ = PIL.Image.open(UpperCAmelCase__) else: A__ = path.split('''::''')[-1] try: A__ = string_to_dict(UpperCAmelCase__ , config.HUB_DATASETS_URL)['''repo_id'''] A__ = token_per_repo_id.get(UpperCAmelCase__) except ValueError: A__ = None with xopen(UpperCAmelCase__ , '''rb''' , use_auth_token=UpperCAmelCase__) as f: A__ = BytesIO(f.read()) A__ = PIL.Image.open(bytes_) else: A__ = PIL.Image.open(BytesIO(bytes_)) image.load() # to avoid "Too many open files" errors return image def SCREAMING_SNAKE_CASE ( self : Dict) ->Union["FeatureType", Dict[str, "FeatureType"]]: '''simple docstring''' from .features import Value return ( self if self.decode else { "bytes": Value('''binary'''), "path": Value('''string'''), } ) def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : Union[pa.StringArray, pa.StructArray, pa.ListArray]) ->pa.StructArray: '''simple docstring''' if pa.types.is_string(storage.type): A__ = pa.array([None] * len(UpperCAmelCase__) , type=pa.binary()) A__ = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null()) elif pa.types.is_binary(storage.type): A__ = pa.array([None] * len(UpperCAmelCase__) , type=pa.string()) A__ = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null()) elif pa.types.is_struct(storage.type): if storage.type.get_field_index('''bytes''') >= 0: A__ = storage.field('''bytes''') else: A__ = pa.array([None] * len(UpperCAmelCase__) , type=pa.binary()) if storage.type.get_field_index('''path''') >= 0: A__ = storage.field('''path''') else: A__ = pa.array([None] * len(UpperCAmelCase__) , type=pa.string()) A__ = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null()) elif pa.types.is_list(storage.type): A__ = pa.array( [encode_np_array(np.array(UpperCAmelCase__))['''bytes'''] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) A__ = pa.array([None] * len(UpperCAmelCase__) , type=pa.string()) A__ = pa.StructArray.from_arrays( [bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null()) return array_cast(UpperCAmelCase__ , self.pa_type) def SCREAMING_SNAKE_CASE ( self : List[Any] , UpperCAmelCase__ : pa.StructArray) ->pa.StructArray: '''simple docstring''' @no_op_if_value_is_null def path_to_bytes(UpperCAmelCase__ : Dict): with xopen(UpperCAmelCase__ , '''rb''') as f: A__ = f.read() return bytes_ A__ = pa.array( [ (path_to_bytes(x['''path''']) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) A__ = pa.array( [os.path.basename(UpperCAmelCase__) if path is not None else None for path in storage.field('''path''').to_pylist()] , type=pa.string() , ) A__ = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null()) return array_cast(UpperCAmelCase__ , self.pa_type) def SCREAMING_SNAKE_CASE ( ) -> List[str]: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() A__ = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def SCREAMING_SNAKE_CASE ( lowercase_ ) -> bytes: """simple docstring""" A__ = BytesIO() if image.format in list_image_compression_formats(): A__ = image.format else: A__ = '''PNG''' if image.mode in ['''1''', '''L''', '''LA''', '''RGB''', '''RGBA'''] else '''TIFF''' image.save(lowercase_ , format=lowercase_ ) return buffer.getvalue() def SCREAMING_SNAKE_CASE ( lowercase_ ) -> dict: """simple docstring""" if hasattr(lowercase_ , '''filename''' ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(lowercase_ )} def SCREAMING_SNAKE_CASE ( lowercase_ ) -> dict: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) A__ = array.dtype A__ = dtype.byteorder if dtype.byteorder != '''=''' else _NATIVE_BYTEORDER A__ = dtype.kind A__ = dtype.itemsize A__ = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: A__ = np.dtype('''|u1''' ) if dtype_kind not in ["u", "i"]: raise TypeError( f"""Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.""" ) if dtype is not dest_dtype: warnings.warn(f"""Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'""" ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: A__ = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: A__ = dtype_byteorder + dtype_kind + str(lowercase_ ) A__ = np.dtype(lowercase_ ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(f"""Downcasting array dtype {dtype} to {dest_dtype} to be compatible with 'Pillow'""" ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( f"""Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}""" ) A__ = PIL.Image.fromarray(array.astype(lowercase_ ) ) return {"path": None, "bytes": image_to_bytes(lowercase_ )} def SCREAMING_SNAKE_CASE ( lowercase_ ) -> List[dict]: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('''To support encoding images, please install \'Pillow\'.''' ) if objs: A__ , A__ = first_non_null_value(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(lowercase_ , np.ndarray ): A__ = no_op_if_value_is_null(lowercase_ ) return [obj_to_image_dict_func(lowercase_ ) for obj in objs] elif isinstance(lowercase_ , PIL.Image.Image ): A__ = no_op_if_value_is_null(lowercase_ ) return [obj_to_image_dict_func(lowercase_ ) for obj in objs] else: return objs else: return objs
87
def UpperCamelCase_( lowerCamelCase_ ) -> int: if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: _lowercase : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Tuple = 0 _lowercase : List[str] = 2 while digits < n: index += 1 _lowercase : Optional[int] = len(str(fibonacci(lowerCamelCase_ ) ) ) return index def UpperCamelCase_( lowerCamelCase_ = 1000 ) -> int: return fibonacci_digits_index(lowerCamelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
89
0
"""simple docstring""" import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase = 256 class lowercase__ ( A_ ): __UpperCAmelCase = ['''melgan'''] def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> None: super().__init__() # From MELGAN _lowerCamelCase : Dict = math.log(1e-5) # Matches MelGAN training. _lowerCamelCase : List[Any] = 4.0 # Largest value for most examples _lowerCamelCase : Dict = 128 self.register_modules( notes_encoder=SCREAMING_SNAKE_CASE , continuous_encoder=SCREAMING_SNAKE_CASE , decoder=SCREAMING_SNAKE_CASE , scheduler=SCREAMING_SNAKE_CASE , melgan=SCREAMING_SNAKE_CASE , ) def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=(-1.0, 1.0) , SCREAMING_SNAKE_CASE=False) -> Any: _lowerCamelCase , _lowerCamelCase : Union[str, Any] = output_range if clip: _lowerCamelCase : Optional[int] = torch.clip(SCREAMING_SNAKE_CASE , self.min_value , self.max_value) # Scale to [0, 1]. _lowerCamelCase : Optional[int] = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=(-1.0, 1.0) , SCREAMING_SNAKE_CASE=False) -> List[str]: _lowerCamelCase , _lowerCamelCase : str = input_range _lowerCamelCase : List[Any] = torch.clip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) if clip else outputs # Scale to [0, 1]. _lowerCamelCase : str = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) -> Union[str, Any]: _lowerCamelCase : Tuple = input_tokens > 0 _lowerCamelCase , _lowerCamelCase : str = self.notes_encoder( encoder_input_tokens=SCREAMING_SNAKE_CASE , encoder_inputs_mask=SCREAMING_SNAKE_CASE) _lowerCamelCase , _lowerCamelCase : Tuple = self.continuous_encoder( encoder_inputs=SCREAMING_SNAKE_CASE , encoder_inputs_mask=SCREAMING_SNAKE_CASE) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def UpperCamelCase_ ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) -> Union[str, Any]: _lowerCamelCase : Optional[int] = noise_time if not torch.is_tensor(SCREAMING_SNAKE_CASE): _lowerCamelCase : int = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device) elif torch.is_tensor(SCREAMING_SNAKE_CASE) and len(timesteps.shape) == 0: _lowerCamelCase : Tuple = timesteps[None].to(input_tokens.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML _lowerCamelCase : Union[str, Any] = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device) _lowerCamelCase : List[str] = self.decoder( encodings_and_masks=SCREAMING_SNAKE_CASE , decoder_input_tokens=SCREAMING_SNAKE_CASE , decoder_noise_time=SCREAMING_SNAKE_CASE) return logits @torch.no_grad() def __call__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = 100 , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = "numpy" , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = 1 , ) -> Union[AudioPipelineOutput, Tuple]: if (callback_steps is None) or ( callback_steps is not None and (not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) or callback_steps <= 0) ): raise ValueError( F'`callback_steps` has to be a positive integer but is {callback_steps} of type' F' {type(SCREAMING_SNAKE_CASE)}.') _lowerCamelCase : Optional[Any] = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa) _lowerCamelCase : Dict = np.zeros([1, 0, self.n_dims] , np.floataa) _lowerCamelCase : List[Any] = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=SCREAMING_SNAKE_CASE , device=self.device) for i, encoder_input_tokens in enumerate(SCREAMING_SNAKE_CASE): if i == 0: _lowerCamelCase : Any = torch.from_numpy(pred_mel[:1].copy()).to( device=self.device , dtype=self.decoder.dtype) # The first chunk has no previous context. _lowerCamelCase : Optional[int] = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=SCREAMING_SNAKE_CASE , device=self.device) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. _lowerCamelCase : Dict = ones _lowerCamelCase : List[Any] = self.scale_features( SCREAMING_SNAKE_CASE , output_range=[-1.0, 1.0] , clip=SCREAMING_SNAKE_CASE) _lowerCamelCase : Dict = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens]).to(device=self.device) , continuous_inputs=SCREAMING_SNAKE_CASE , continuous_mask=SCREAMING_SNAKE_CASE , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop _lowerCamelCase : List[str] = randn_tensor( shape=encoder_continuous_inputs.shape , generator=SCREAMING_SNAKE_CASE , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps)): _lowerCamelCase : str = self.decode( encodings_and_masks=SCREAMING_SNAKE_CASE , input_tokens=SCREAMING_SNAKE_CASE , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 _lowerCamelCase : Tuple = self.scheduler.step(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE).prev_sample _lowerCamelCase : Any = self.scale_to_features(SCREAMING_SNAKE_CASE , input_range=[-1.0, 1.0]) _lowerCamelCase : int = mel[:1] _lowerCamelCase : Optional[int] = mel.cpu().float().numpy() _lowerCamelCase : str = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) logger.info("""Generated segment""" , SCREAMING_SNAKE_CASE) if output_type == "numpy" and not is_onnx_available(): raise ValueError( """Cannot return output in 'np' format if ONNX is not available. Make sure to have ONNX installed or set 'output_type' to 'mel'.""") elif output_type == "numpy" and self.melgan is None: raise ValueError( """Cannot return output in 'np' format if melgan component is not defined. Make sure to define `self.melgan` or set 'output_type' to 'mel'.""") if output_type == "numpy": _lowerCamelCase : Union[str, Any] = self.melgan(input_features=full_pred_mel.astype(np.floataa)) else: _lowerCamelCase : Optional[Any] = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=SCREAMING_SNAKE_CASE)
88
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Union[str, Any] = ["ConditionalDetrFeatureExtractor"] SCREAMING_SNAKE_CASE : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Any = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
90
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function SCREAMING_SNAKE_CASE : Union[str, Any] = 1.0_5457_1817E-34 # unit of ℏ : J * s SCREAMING_SNAKE_CASE : int = 3E8 # unit of c : m * s^-1 def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> dict[str, float]: if (force, area, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if force < 0: raise ValueError('Magnitude of force can not be negative' ) if distance < 0: raise ValueError('Distance can not be negative' ) if area < 0: raise ValueError('Area can not be negative' ) if force == 0: _lowercase : int = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: _lowercase : List[Any] = (240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: _lowercase : List[Any] = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError('One and only one argument must be 0' ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
89
0
"""simple docstring""" class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Optional[int] ) -> List[str]: A = 0 A = 0 A = {} def _SCREAMING_SNAKE_CASE ( self : int ,A_ : Optional[Any] ) -> int: if vertex not in self.adjacency: A = {} self.num_vertices += 1 def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ,A_ : Tuple ,A_ : Any ,A_ : str ) -> Optional[Any]: self.add_vertex(A_ ) self.add_vertex(A_ ) if head == tail: return A = weight A = weight def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: A = self.get_edges() for edge in edges: A , A , A = edge edges.remove((tail, head, weight) ) for i in range(len(A_ ) ): A = list(edges[i] ) edges.sort(key=lambda A_ : e[2] ) for i in range(len(A_ ) - 1 ): if edges[i][2] >= edges[i + 1][2]: A = edges[i][2] + 1 for edge in edges: A , A , A = edge A = weight A = weight def __str__( self : Union[str, Any] ) -> str: A = '' for tail in self.adjacency: for head in self.adjacency[tail]: A = self.adjacency[head][tail] string += F'{head} -> {tail} == {weight}\n' return string.rstrip('\n' ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> int: A = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> int: return self.adjacency.keys() @staticmethod def _SCREAMING_SNAKE_CASE ( A_ : Dict=None ,A_ : int=None ) -> Tuple: A = Graph() if vertices is None: A = [] if edges is None: A = [] for vertex in vertices: g.add_vertex(A_ ) for edge in edges: g.add_edge(*A_ ) return g class lowerCAmelCase_ : '''simple docstring''' def __init__( self : str ) -> Union[str, Any]: A = {} A = {} def __len__( self : int ) -> int: return len(self.parent ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ,A_ : Any ) -> Any: if item in self.parent: return self.find(A_ ) A = item A = 0 return item def _SCREAMING_SNAKE_CASE ( self : Any ,A_ : Tuple ) -> Tuple: if item not in self.parent: return self.make_set(A_ ) if item != self.parent[item]: A = self.find(self.parent[item] ) return self.parent[item] def _SCREAMING_SNAKE_CASE ( self : Optional[int] ,A_ : Dict ,A_ : Optional[Any] ) -> List[str]: A = self.find(A_ ) A = self.find(A_ ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: A = roota return roota if self.rank[roota] < self.rank[roota]: A = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 A = roota return roota return None @staticmethod def _SCREAMING_SNAKE_CASE ( A_ : Any ) -> Optional[int]: A = graph.num_vertices A = Graph.UnionFind() A = [] while num_components > 1: A = {} for vertex in graph.get_vertices(): A = -1 A = graph.get_edges() for edge in edges: A , A , A = edge edges.remove((tail, head, weight) ) for edge in edges: A , A , A = edge A = union_find.find(A_ ) A = union_find.find(A_ ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: A = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: A = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: A , A , A = cheap_edge[vertex] if union_find.find(A_ ) != union_find.find(A_ ): union_find.union(A_ ,A_ ) mst_edges.append(cheap_edge[vertex] ) A = num_components - 1 A = Graph.build(edges=A_ ) return mst
91
def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) _lowercase : List[str] = 0 _lowercase : Optional[int] = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Any = [int(lowerCamelCase_ ) for i in num_string] _lowercase : List[Any] = 1 for i in range(0 , len(lowerCamelCase_ ) ): total *= numbers[i] _lowercase : Optional[Any] = str(lowerCamelCase_ ) steps += 1 return steps def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) _lowercase : Optional[int] = 0 _lowercase : str = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Dict = [int(lowerCamelCase_ ) for i in num_string] _lowercase : Any = 0 for i in range(0 , len(lowerCamelCase_ ) ): total += numbers[i] _lowercase : Dict = str(lowerCamelCase_ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
89
0
'''simple docstring''' def _lowerCAmelCase ( __magic_name__ : int ) -> int: if not isinstance(__magic_name__ , __magic_name__ ): lowercase : List[str] =f'''Input value of [number={number}] must be an integer''' raise TypeError(__magic_name__ ) if number < 1: lowercase : List[str] =f'''Input value of [number={number}] must be > 0''' raise ValueError(__magic_name__ ) lowercase : Dict =1 for i in range(1 , __magic_name__ ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
92
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: # initialize config if "resnet-50" in model_name: _lowercase : Union[str, Any] = ResNetConfig.from_pretrained('microsoft/resnet-50' ) elif "resnet-101" in model_name: _lowercase : Optional[Any] = ResNetConfig.from_pretrained('microsoft/resnet-101' ) else: raise ValueError('Model name should include either resnet50 or resnet101' ) _lowercase : Tuple = DetrConfig(use_timm_backbone=lowerCamelCase_ , backbone_config=lowerCamelCase_ ) # set label attributes _lowercase : Any = 'panoptic' in model_name if is_panoptic: _lowercase : List[Any] = 250 else: _lowercase : str = 91 _lowercase : List[Any] = 'huggingface/label-files' _lowercase : Any = 'coco-detection-id2label.json' _lowercase : Tuple = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : int = idalabel _lowercase : Any = {v: k for k, v in idalabel.items()} return config, is_panoptic def UpperCamelCase_( lowerCamelCase_ ) -> Any: # here we list all keys to be renamed (original name on the left, our name on the right) _lowercase : List[str] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.conv1.weight', 'backbone.conv_encoder.model.embedder.embedder.convolution.weight') ) rename_keys.append(('backbone.0.body.bn1.weight', 'backbone.conv_encoder.model.embedder.embedder.normalization.weight') ) rename_keys.append(('backbone.0.body.bn1.bias', 'backbone.conv_encoder.model.embedder.embedder.normalization.bias') ) rename_keys.append(('backbone.0.body.bn1.running_mean', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_mean') ) rename_keys.append(('backbone.0.body.bn1.running_var', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_var') ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var''', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var''', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ] ) return rename_keys def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : str = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> str: _lowercase : Any = '' if is_panoptic: _lowercase : Optional[Any] = 'detr.' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowercase : int = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Tuple = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[str] = in_proj_weight[:256, :] _lowercase : Tuple = in_proj_bias[:256] _lowercase : List[Any] = in_proj_weight[256:512, :] _lowercase : Any = in_proj_bias[256:512] _lowercase : int = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _lowercase : str = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Optional[int] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : Union[str, Any] = in_proj_weight[:256, :] _lowercase : Dict = in_proj_bias[:256] _lowercase : Tuple = in_proj_weight[256:512, :] _lowercase : Dict = in_proj_bias[256:512] _lowercase : str = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _lowercase : Tuple = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) _lowercase : Dict = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _lowercase : List[str] = in_proj_weight_cross_attn[:256, :] _lowercase : Tuple = in_proj_bias_cross_attn[:256] _lowercase : str = in_proj_weight_cross_attn[256:512, :] _lowercase : Union[str, Any] = in_proj_bias_cross_attn[256:512] _lowercase : List[Any] = in_proj_weight_cross_attn[-256:, :] _lowercase : Dict = in_proj_bias_cross_attn[-256:] def UpperCamelCase_( ) -> List[Any]: _lowercase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : str = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) return im @torch.no_grad() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=False ) -> List[Any]: _lowercase , _lowercase : int = get_detr_config(lowerCamelCase_ ) # load original model from torch hub _lowercase : int = { 'detr-resnet-50': 'detr_resnet50', 'detr-resnet-101': 'detr_resnet101', } logger.info(F'''Converting model {model_name}...''' ) _lowercase : Optional[Any] = torch.hub.load('facebookresearch/detr' , model_name_to_original_name[model_name] , pretrained=lowerCamelCase_ ).eval() _lowercase : str = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCamelCase_ ): if is_panoptic: _lowercase : str = 'detr.' + src rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCamelCase_ , is_panoptic=lowerCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowercase : List[Any] = 'detr.model.' if is_panoptic else 'model.' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('detr' ) and not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ) ): _lowercase : Tuple = state_dict.pop(lowerCamelCase_ ) _lowercase : int = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val elif key.startswith('bbox_attention' ) or key.startswith('mask_head' ): continue else: _lowercase : Optional[Any] = state_dict.pop(lowerCamelCase_ ) _lowercase : Union[str, Any] = val else: if not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : List[str] = val # finally, create HuggingFace model and load state dict _lowercase : Optional[Any] = DetrForSegmentation(lowerCamelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCamelCase_ ) model.load_state_dict(lowerCamelCase_ ) model.eval() # verify our conversion on an image _lowercase : str = 'coco_panoptic' if is_panoptic else 'coco_detection' _lowercase : Optional[int] = DetrImageProcessor(format=lowerCamelCase_ ) _lowercase : str = processor(images=prepare_img() , return_tensors='pt' ) _lowercase : Tuple = encoding['pixel_values'] _lowercase : int = detr(lowerCamelCase_ ) _lowercase : Tuple = model(lowerCamelCase_ ) assert torch.allclose(outputs.logits , original_outputs['pred_logits'] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs['pred_boxes'] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['pred_masks'] , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info('Uploading PyTorch model and image processor to the hub...' ) model.push_to_hub(F'''nielsr/{model_name}''' ) processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_name", default="detr-resnet-50", type=str, choices=["detr-resnet-50", "detr-resnet-101"], help="Name of the DETR model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the model to the hub or not.") SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
"""simple docstring""" import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE ) as metadata_file: lowerCAmelCase__ :Union[str, Any] = json.load(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[str] = LukeConfig(use_entity_aware_attention=_SCREAMING_SNAKE_CASE , **metadata['model_config'] ) # Load in the weights from the checkpoint_path lowerCAmelCase__ :Optional[int] = torch.load(_SCREAMING_SNAKE_CASE , map_location='cpu' )['module'] # Load the entity vocab file lowerCAmelCase__ :Any = load_original_entity_vocab(_SCREAMING_SNAKE_CASE ) # add an entry for [MASK2] lowerCAmelCase__ :Union[str, Any] = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 lowerCAmelCase__ :List[str] = XLMRobertaTokenizer.from_pretrained(metadata['model_config']['bert_model_name'] ) # Add special tokens to the token vocabulary for downstream tasks lowerCAmelCase__ :List[Any] = AddedToken('<ent>' , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Optional[Any] = AddedToken('<ent2>' , lstrip=_SCREAMING_SNAKE_CASE , rstrip=_SCREAMING_SNAKE_CASE ) tokenizer.add_special_tokens({'additional_special_tokens': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"Saving tokenizer to {pytorch_dump_folder_path}" ) tokenizer.save_pretrained(_SCREAMING_SNAKE_CASE ) with open(os.path.join(_SCREAMING_SNAKE_CASE , 'tokenizer_config.json' ) , 'r' ) as f: lowerCAmelCase__ :Optional[Any] = json.load(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Tuple = 'MLukeTokenizer' with open(os.path.join(_SCREAMING_SNAKE_CASE , 'tokenizer_config.json' ) , 'w' ) as f: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) with open(os.path.join(_SCREAMING_SNAKE_CASE , MLukeTokenizer.vocab_files_names['entity_vocab_file'] ) , 'w' ) as f: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Union[str, Any] = MLukeTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ) # Initialize the embeddings of the special tokens lowerCAmelCase__ :str = tokenizer.convert_tokens_to_ids(['@'] )[0] lowerCAmelCase__ :int = tokenizer.convert_tokens_to_ids(['#'] )[0] lowerCAmelCase__ :Dict = state_dict['embeddings.word_embeddings.weight'] lowerCAmelCase__ :Union[str, Any] = word_emb[ent_init_index].unsqueeze(0 ) lowerCAmelCase__ :str = word_emb[enta_init_index].unsqueeze(0 ) lowerCAmelCase__ :Any = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: lowerCAmelCase__ :str = state_dict[bias_name] lowerCAmelCase__ :List[str] = decoder_bias[ent_init_index].unsqueeze(0 ) lowerCAmelCase__ :str = decoder_bias[enta_init_index].unsqueeze(0 ) lowerCAmelCase__ :Union[str, Any] = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowerCAmelCase__ :Dict = F"encoder.layer.{layer_index}.attention.self." lowerCAmelCase__ :int = state_dict[prefix + matrix_name] lowerCAmelCase__ :str = state_dict[prefix + matrix_name] lowerCAmelCase__ :str = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowerCAmelCase__ :Optional[Any] = state_dict['entity_embeddings.entity_embeddings.weight'] lowerCAmelCase__ :Union[str, Any] = entity_emb[entity_vocab['[MASK]']].unsqueeze(0 ) lowerCAmelCase__ :Optional[Any] = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' lowerCAmelCase__ :List[Any] = state_dict['entity_predictions.bias'] lowerCAmelCase__ :Dict = entity_prediction_bias[entity_vocab['[MASK]']].unsqueeze(0 ) lowerCAmelCase__ :Any = torch.cat([entity_prediction_bias, entity_mask_bias] ) lowerCAmelCase__ :int = LukeForMaskedLM(config=_SCREAMING_SNAKE_CASE ).eval() state_dict.pop('entity_predictions.decoder.weight' ) state_dict.pop('lm_head.decoder.weight' ) state_dict.pop('lm_head.decoder.bias' ) lowerCAmelCase__ :Tuple = OrderedDict() for key, value in state_dict.items(): if not (key.startswith('lm_head' ) or key.startswith('entity_predictions' )): lowerCAmelCase__ :List[str] = state_dict[key] else: lowerCAmelCase__ :Union[str, Any] = state_dict[key] lowerCAmelCase__ , lowerCAmelCase__ :int = model.load_state_dict(_SCREAMING_SNAKE_CASE , strict=_SCREAMING_SNAKE_CASE ) if set(_SCREAMING_SNAKE_CASE ) != {"luke.embeddings.position_ids"}: raise ValueError(F"Unexpected unexpected_keys: {unexpected_keys}" ) if set(_SCREAMING_SNAKE_CASE ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(F"Unexpected missing_keys: {missing_keys}" ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs lowerCAmelCase__ :List[Any] = MLukeTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE , task='entity_classification' ) lowerCAmelCase__ :Union[str, Any] = 'ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).' lowerCAmelCase__ :Optional[Any] = (0, 9) lowerCAmelCase__ :int = tokenizer(_SCREAMING_SNAKE_CASE , entity_spans=[span] , return_tensors='pt' ) lowerCAmelCase__ :List[str] = model(**_SCREAMING_SNAKE_CASE ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base lowerCAmelCase__ :Union[str, Any] = torch.Size((1, 33, 768) ) lowerCAmelCase__ :Optional[int] = torch.tensor([[0.0_8_9_2, 0.0_5_9_6, -0.2_8_1_9], [0.0_1_3_4, 0.1_1_9_9, 0.0_5_7_3], [-0.0_1_6_9, 0.0_9_2_7, 0.0_6_4_4]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base lowerCAmelCase__ :Optional[Any] = torch.Size((1, 1, 768) ) lowerCAmelCase__ :Any = torch.tensor([[-0.1_4_8_2, 0.0_6_0_9, 0.0_3_2_2]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( F"Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is" F" {expected_shape}" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ): raise ValueError # Verify masked word/entity prediction lowerCAmelCase__ :str = MLukeTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Tuple = 'Tokyo is the capital of <mask>.' lowerCAmelCase__ :Optional[int] = (24, 30) lowerCAmelCase__ :Any = tokenizer(_SCREAMING_SNAKE_CASE , entity_spans=[span] , return_tensors='pt' ) lowerCAmelCase__ :int = model(**_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Optional[int] = encoding['input_ids'][0].tolist() lowerCAmelCase__ :Optional[int] = input_ids.index(tokenizer.convert_tokens_to_ids('<mask>' ) ) lowerCAmelCase__ :str = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :Optional[int] = outputs.entity_logits[0][0].argmax().item() lowerCAmelCase__ :Dict = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith('en:' )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print('Saving PyTorch model to {}'.format(_SCREAMING_SNAKE_CASE ) ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) def __A (_SCREAMING_SNAKE_CASE ) ->str: """simple docstring""" lowerCAmelCase__ :List[Any] = ['[MASK]', '[PAD]', '[UNK]'] lowerCAmelCase__ :Any = [json.loads(_SCREAMING_SNAKE_CASE ) for line in open(_SCREAMING_SNAKE_CASE )] lowerCAmelCase__ :Optional[Any] = {} for entry in data: lowerCAmelCase__ :Optional[int] = entry['id'] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: lowerCAmelCase__ :Tuple = entity_id break lowerCAmelCase__ :int = F"{language}:{entity_name}" lowerCAmelCase__ :int = entity_id return new_mapping if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument("""--checkpoint_path""", type=str, help="""Path to a pytorch_model.bin file.""") parser.add_argument( """--metadata_path""", default=None, type=str, help="""Path to a metadata.json file, defining the configuration.""" ) parser.add_argument( """--entity_vocab_path""", default=None, type=str, help="""Path to an entity_vocab.tsv file, containing the entity vocabulary.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to where to dump the output PyTorch model.""" ) parser.add_argument( """--model_size""", default="""base""", type=str, choices=["""base""", """large"""], help="""Size of the model to be converted.""" ) __A = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
93
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput SCREAMING_SNAKE_CASE : str = "scheduler_config.json" class _lowerCamelCase( _a ): lowercase_ : Any = 1 lowercase_ : Dict = 2 lowercase_ : Union[str, Any] = 3 lowercase_ : Tuple = 4 lowercase_ : Optional[Any] = 5 @dataclass class _lowerCamelCase( _a ): lowercase_ : jnp.ndarray class _lowerCamelCase: lowercase_ : Union[str, Any] = SCHEDULER_CONFIG_NAME lowercase_ : str = ["""dtype"""] lowercase_ : Dict = [] lowercase_ : int = True @classmethod def UpperCamelCase ( cls, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase=False, **lowerCamelCase, ) -> Optional[int]: """simple docstring""" _lowercase , _lowercase : Optional[int] = cls.load_config( pretrained_model_name_or_path=lowerCamelCase, subfolder=lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase, ) _lowercase , _lowercase : Tuple = cls.from_config(lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase) if hasattr(lowerCamelCase, 'create_state') and getattr(lowerCamelCase, 'has_state', lowerCamelCase): _lowercase : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, **lowerCamelCase) -> Any: """simple docstring""" self.save_config(save_directory=lowerCamelCase, push_to_hub=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return self._get_compatibles() @classmethod def UpperCamelCase ( cls) -> Any: """simple docstring""" _lowercase : Any = list(set([cls.__name__] + cls._compatibles)) _lowercase : Dict = importlib.import_module(__name__.split('.')[0]) _lowercase : Any = [ getattr(lowerCamelCase, lowerCamelCase) for c in compatible_classes_str if hasattr(lowerCamelCase, lowerCamelCase) ] return compatible_classes def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> jnp.ndarray: assert len(lowerCamelCase_ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(lowerCamelCase_ ) - x.ndim) ) , lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=0.9_99 , lowerCamelCase_=jnp.floataa ) -> jnp.ndarray: def alpha_bar(lowerCamelCase_ ): return math.cos((time_step + 0.0_08) / 1.0_08 * math.pi / 2 ) ** 2 _lowercase : List[Any] = [] for i in range(lowerCamelCase_ ): _lowercase : Any = i / num_diffusion_timesteps _lowercase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(lowerCamelCase_ ) / alpha_bar(lowerCamelCase_ ) , lowerCamelCase_ ) ) return jnp.array(lowerCamelCase_ , dtype=lowerCamelCase_ ) @flax.struct.dataclass class _lowerCamelCase: lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray @classmethod def UpperCamelCase ( cls, lowerCamelCase) -> str: """simple docstring""" _lowercase : int = scheduler.config if config.trained_betas is not None: _lowercase : str = jnp.asarray(config.trained_betas, dtype=scheduler.dtype) elif config.beta_schedule == "linear": _lowercase : List[Any] = jnp.linspace(config.beta_start, config.beta_end, config.num_train_timesteps, dtype=scheduler.dtype) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _lowercase : Dict = ( jnp.linspace( config.beta_start**0.5, config.beta_end**0.5, config.num_train_timesteps, dtype=scheduler.dtype) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _lowercase : Optional[int] = betas_for_alpha_bar(config.num_train_timesteps, dtype=scheduler.dtype) else: raise NotImplementedError( F'''beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}''') _lowercase : List[str] = 1.0 - betas _lowercase : Union[str, Any] = jnp.cumprod(lowerCamelCase, axis=0) return cls( alphas=lowerCamelCase, betas=lowerCamelCase, alphas_cumprod=lowerCamelCase, ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : str = state.alphas_cumprod _lowercase : str = alphas_cumprod[timesteps] ** 0.5 _lowercase : Optional[Any] = sqrt_alpha_prod.flatten() _lowercase : Tuple = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) _lowercase : Union[str, Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowercase : Optional[Any] = sqrt_one_minus_alpha_prod.flatten() _lowercase : int = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase , _lowercase : Optional[int] = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: _lowercase , _lowercase : Tuple = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
89
0
'''simple docstring''' import cmath import math def lowercase_ ( __A : float , __A : float , __A : float , __A : float ) -> complex: """simple docstring""" lowercase : Any =math.radians(__A ) lowercase : Union[str, Any] =math.radians(__A ) # Convert voltage and current to rectangular form lowercase : Tuple =cmath.rect(__A , __A ) lowercase : List[str] =cmath.rect(__A , __A ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
94
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> float: if not nums: raise ValueError('List is empty' ) return sum(lowerCamelCase_ ) / len(lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
89
0
"""simple docstring""" class UpperCamelCase_ : def __init__( self : List[Any] , lowerCAmelCase_ : list[int] ) -> None: UpperCAmelCase_ : str = len(lowerCAmelCase_ ) UpperCAmelCase_ : Dict = [0] * len_array if len_array > 0: UpperCAmelCase_ : str = array[0] for i in range(1 , lowerCAmelCase_ ): UpperCAmelCase_ : int = self.prefix_sum[i - 1] + array[i] def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> int: if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase_ : int ) -> bool: UpperCAmelCase_ : int = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(lowerCAmelCase_ ) return False if __name__ == "__main__": import doctest doctest.testmod()
95
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def UpperCamelCase_( ) -> List[Any]: _lowercase : int = ArgumentParser('Transformers CLI tool' , usage='transformers-cli <command> [<args>]' ) _lowercase : Optional[Any] = parser.add_subparsers(help='transformers-cli command helpers' ) # Register commands ConvertCommand.register_subcommand(lowerCamelCase_ ) DownloadCommand.register_subcommand(lowerCamelCase_ ) EnvironmentCommand.register_subcommand(lowerCamelCase_ ) RunCommand.register_subcommand(lowerCamelCase_ ) ServeCommand.register_subcommand(lowerCamelCase_ ) UserCommands.register_subcommand(lowerCamelCase_ ) AddNewModelCommand.register_subcommand(lowerCamelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCamelCase_ ) LfsCommands.register_subcommand(lowerCamelCase_ ) PTtoTFCommand.register_subcommand(lowerCamelCase_ ) # Let's go _lowercase : Any = parser.parse_args() if not hasattr(lowerCamelCase_ , 'func' ): parser.print_help() exit(1 ) # Run _lowercase : Optional[int] = args.func(lowerCamelCase_ ) service.run() if __name__ == "__main__": main()
89
0
"""simple docstring""" import gc import unittest from transformers import CTRLConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( CTRL_PRETRAINED_MODEL_ARCHIVE_LIST, CTRLForSequenceClassification, CTRLLMHeadModel, CTRLModel, ) class __A : def __init__( self : Optional[int] , __snake_case : int , __snake_case : str=1_4 , __snake_case : Any=7 , __snake_case : Dict=True , __snake_case : Union[str, Any]=True , __snake_case : List[str]=True , __snake_case : Any=True , __snake_case : Dict=True , __snake_case : int=9_9 , __snake_case : Tuple=3_2 , __snake_case : List[str]=5 , __snake_case : Optional[Any]=4 , __snake_case : Optional[int]=3_7 , __snake_case : Optional[int]="gelu" , __snake_case : Tuple=0.1 , __snake_case : Optional[int]=0.1 , __snake_case : Optional[int]=5_1_2 , __snake_case : List[Any]=1_6 , __snake_case : Tuple=2 , __snake_case : str=0.02 , __snake_case : str=3 , __snake_case : int=4 , __snake_case : Dict=None , ) -> List[str]: __magic_name__: Tuple = parent __magic_name__: int = batch_size __magic_name__: str = seq_length __magic_name__: Optional[int] = is_training __magic_name__: Dict = use_token_type_ids __magic_name__: str = use_input_mask __magic_name__: int = use_labels __magic_name__: Union[str, Any] = use_mc_token_ids __magic_name__: str = vocab_size __magic_name__: List[str] = hidden_size __magic_name__: Any = num_hidden_layers __magic_name__: int = num_attention_heads __magic_name__: Dict = intermediate_size __magic_name__: Optional[Any] = hidden_act __magic_name__: Any = hidden_dropout_prob __magic_name__: Dict = attention_probs_dropout_prob __magic_name__: List[Any] = max_position_embeddings __magic_name__: List[str] = type_vocab_size __magic_name__: int = type_sequence_label_size __magic_name__: Optional[int] = initializer_range __magic_name__: Optional[int] = num_labels __magic_name__: str = num_choices __magic_name__: Optional[Any] = scope __magic_name__: Union[str, Any] = self.vocab_size - 1 def lowerCamelCase__ ( self : Optional[int] ) -> List[Any]: __magic_name__: Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__: Dict = None if self.use_input_mask: __magic_name__: Any = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__: List[Any] = None if self.use_token_type_ids: __magic_name__: List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__: Dict = None if self.use_mc_token_ids: __magic_name__: Any = ids_tensor([self.batch_size, self.num_choices] , self.seq_length ) __magic_name__: List[Any] = None __magic_name__: Any = None __magic_name__: List[Any] = None if self.use_labels: __magic_name__: List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__: Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __magic_name__: Dict = ids_tensor([self.batch_size] , self.num_choices ) __magic_name__: Tuple = self.get_config() __magic_name__: Optional[int] = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, token_type_ids, mc_token_ids, sequence_labels, token_labels, choice_labels, ) def lowerCamelCase__ ( self : Any ) -> Optional[Any]: return CTRLConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) def lowerCamelCase__ ( self : Any , __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : List[Any] , __snake_case : Optional[int] , *__snake_case : Dict ) -> Any: __magic_name__: Any = CTRLModel(config=__snake_case ) model.to(__snake_case ) model.eval() model(__snake_case , token_type_ids=__snake_case , head_mask=__snake_case ) model(__snake_case , token_type_ids=__snake_case ) __magic_name__: List[Any] = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(len(result.past_key_values ) , config.n_layer ) def lowerCamelCase__ ( self : Union[str, Any] , __snake_case : str , __snake_case : int , __snake_case : Any , __snake_case : Optional[Any] , __snake_case : Dict , *__snake_case : Any ) -> Union[str, Any]: __magic_name__: List[Any] = CTRLLMHeadModel(__snake_case ) model.to(__snake_case ) model.eval() __magic_name__: Optional[Any] = model(__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self : Dict ) -> Optional[int]: __magic_name__: List[str] = self.prepare_config_and_inputs() ( ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ( __magic_name__ ), ): int = config_and_inputs __magic_name__: int = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask} return config, inputs_dict def lowerCamelCase__ ( self : List[str] , __snake_case : List[Any] , __snake_case : str , __snake_case : str , __snake_case : Optional[int] , *__snake_case : Union[str, Any] ) -> str: __magic_name__: int = self.num_labels __magic_name__: int = CTRLForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() __magic_name__: str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__: int = model(__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) @require_torch class __A ( SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ ,unittest.TestCase ): UpperCAmelCase__ = (CTRLModel, CTRLLMHeadModel, CTRLForSequenceClassification) if is_torch_available() else () UpperCAmelCase__ = (CTRLLMHeadModel,) if is_torch_available() else () UpperCAmelCase__ = ( { "feature-extraction": CTRLModel, "text-classification": CTRLForSequenceClassification, "text-generation": CTRLLMHeadModel, "zero-shot": CTRLForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = False def lowerCamelCase__ ( self : Optional[Any] , __snake_case : Optional[int] , __snake_case : Tuple , __snake_case : List[str] , __snake_case : int , __snake_case : Optional[int] ) -> Dict: if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `CTRLConfig` was never used in pipeline tests, either because of a missing checkpoint or because a tiny # config could not be created. return True return False def lowerCamelCase__ ( self : Union[str, Any] ) -> Union[str, Any]: __magic_name__: str = CTRLModelTester(self ) __magic_name__: Optional[int] = ConfigTester(self , config_class=__snake_case , n_embd=3_7 ) def lowerCamelCase__ ( self : List[str] ) -> Optional[int]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self : Tuple ) -> List[str]: self.config_tester.run_common_tests() def lowerCamelCase__ ( self : List[str] ) -> List[Any]: __magic_name__: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_ctrl_model(*__snake_case ) def lowerCamelCase__ ( self : Any ) -> Tuple: __magic_name__: Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*__snake_case ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCamelCase__ ( self : Tuple ) -> Dict: pass @slow def lowerCamelCase__ ( self : Dict ) -> Optional[Any]: for model_name in CTRL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __magic_name__: Union[str, Any] = CTRLModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip("""The model doesn't support left padding""" ) # and it's not used enough to be worth fixing :) def lowerCamelCase__ ( self : Optional[int] ) -> List[Any]: pass @require_torch class __A ( unittest.TestCase ): def lowerCamelCase__ ( self : Tuple ) -> str: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() torch.cuda.empty_cache() @slow def lowerCamelCase__ ( self : Optional[int] ) -> List[Any]: __magic_name__: Optional[Any] = CTRLLMHeadModel.from_pretrained("""ctrl""" ) model.to(__snake_case ) __magic_name__: List[str] = torch.tensor( [[1_1_8_5_9, 0, 1_6_1_1, 8]] , dtype=torch.long , device=__snake_case ) # Legal the president is __magic_name__: List[Any] = [ 1_1_8_5_9, 0, 1_6_1_1, 8, 5, 1_5_0, 2_6_4_4_9, 2, 1_9, 3_4_8, 4_6_9, 3, 2_5_9_5, 4_8, 2_0_7_4_0, 2_4_6_5_3_3, 2_4_6_5_3_3, 1_9, 3_0, 5, ] # Legal the president is a good guy and I don't want to lose my job. \n \n I have a __magic_name__: Optional[int] = model.generate(__snake_case , do_sample=__snake_case ) self.assertListEqual(output_ids[0].tolist() , __snake_case )
96
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Optional[int] = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def UpperCamelCase ( self, lowerCamelCase=0) -> str: """simple docstring""" _lowercase : Optional[int] = np.random.RandomState(lowerCamelCase) _lowercase : Union[str, Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : List[Any] = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[str] = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Any = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_dummy_inputs() _lowercase : Any = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : Optional[int] = output.images[0, -3:, -3:, -1] _lowercase : int = self.get_dummy_inputs() _lowercase : Union[str, Any] = 3 * [inputs.pop('prompt')] _lowercase : Union[str, Any] = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Tuple = text_inputs['input_ids'] _lowercase : Any = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0] _lowercase : List[Any] = prompt_embeds # forward _lowercase : Union[str, Any] = pipe(**lowerCamelCase) _lowercase : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs() _lowercase : Any = 3 * ['this is a negative prompt'] _lowercase : str = negative_prompt _lowercase : Optional[int] = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : str = output.images[0, -3:, -3:, -1] _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : str = 3 * [inputs.pop('prompt')] _lowercase : Optional[int] = [] for p in [prompt, negative_prompt]: _lowercase : Tuple = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Dict = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0]) _lowercase , _lowercase : str = embeds # forward _lowercase : Dict = pipe(**lowerCamelCase) _lowercase : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = ort.SessionOptions() _lowercase : str = False return options def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'A painting of a squirrel eating a burger' np.random.seed(0) _lowercase : Union[str, Any] = sd_pipe([prompt], guidance_scale=6.0, num_inference_steps=10, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : str = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'open neural network exchange' _lowercase : List[Any] = np.random.RandomState(0) _lowercase : Optional[Any] = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = 'open neural network exchange' _lowercase : str = np.random.RandomState(0) _lowercase : Dict = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[Any] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : List[Any] = 0 def test_callback_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase) -> None: _lowercase : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _lowercase : Any = latents[0, -3:, -3:, -1] _lowercase : Tuple = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _lowercase : List[Any] = latents[0, -3:, -3:, -1] _lowercase : str = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 _lowercase : Any = False _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = 'Andromeda galaxy in a bottle' _lowercase : str = np.random.RandomState(0) pipe( prompt=lowerCamelCase, num_inference_steps=5, guidance_scale=7.5, generator=lowerCamelCase, callback=lowerCamelCase, callback_steps=1, ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) assert isinstance(lowerCamelCase, lowerCamelCase) assert pipe.safety_checker is None _lowercase : Optional[int] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase) _lowercase : Any = OnnxStableDiffusionPipeline.from_pretrained(lowerCamelCase) # sanity check that the pipeline still works assert pipe.safety_checker is None _lowercase : List[str] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None
89
0
import string def a ( snake_case__: str ): '''simple docstring''' lowercase_ = '''''' for i in sequence: lowercase_ = ord(snake_case__ ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def a ( snake_case__: str ): '''simple docstring''' lowercase_ = string.ascii_letters lowercase_ = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(snake_case__ )] if c in letters else c for c in sequence ) def a ( ): '''simple docstring''' from timeit import timeit print('''Running performance benchmarks...''' ) lowercase_ = '''from string import printable ; from __main__ import atbash, atbash_slow''' print(F'''> atbash_slow(): {timeit('atbash_slow(printable)' , setup=snake_case__ )} seconds''' ) print(F'''> atbash(): {timeit('atbash(printable)' , setup=snake_case__ )} seconds''' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(f"{example} encrypted in atbash: {atbash(example)}") benchmark()
97
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : List[Any] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = ["PoolFormerFeatureExtractor"] SCREAMING_SNAKE_CASE : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[Any] = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
89
0
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html lowercase__ : Any = 'platform' import jax import jax.numpy as jnp from transformers.models.blenderbot_small.modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, shift_tokens_right, ) def a__ ( lowercase : Optional[Any], lowercase : List[str], lowercase : Tuple=None, lowercase : Tuple=None, lowercase : int=None, lowercase : Optional[int]=None, lowercase : Any=None, lowercase : Optional[int]=None, ) -> Any: """simple docstring""" if attention_mask is None: _UpperCamelCase = np.where(input_ids != config.pad_token_id, 1, 0 ) if decoder_attention_mask is None: _UpperCamelCase = np.where(decoder_input_ids != config.pad_token_id, 1, 0 ) if head_mask is None: _UpperCamelCase = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _UpperCamelCase = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _UpperCamelCase = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class __lowerCAmelCase : """simple docstring""" def __init__( self : str , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Tuple=13 , lowerCAmelCase__ : Optional[Any]=7 , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : Tuple=False , lowerCAmelCase__ : Dict=99 , lowerCAmelCase__ : Union[str, Any]=16 , lowerCAmelCase__ : Tuple=2 , lowerCAmelCase__ : Dict=4 , lowerCAmelCase__ : str=4 , lowerCAmelCase__ : List[str]="gelu" , lowerCAmelCase__ : str=0.1 , lowerCAmelCase__ : str=0.1 , lowerCAmelCase__ : Any=32 , lowerCAmelCase__ : Dict=2 , lowerCAmelCase__ : List[Any]=1 , lowerCAmelCase__ : Any=0 , lowerCAmelCase__ : Dict=0.02 , ) -> int: '''simple docstring''' _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = seq_length _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = eos_token_id _UpperCamelCase = pad_token_id _UpperCamelCase = bos_token_id _UpperCamelCase = initializer_range def snake_case__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _UpperCamelCase = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _UpperCamelCase = shift_tokens_right(lowerCAmelCase__ , 1 , 2 ) _UpperCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCAmelCase__ , ) _UpperCamelCase = prepare_blenderbot_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return config, inputs_dict def snake_case__ ( self : List[str] ) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.prepare_config_and_inputs() return config, inputs_dict def snake_case__ ( self : Optional[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : Any , lowerCAmelCase__ : List[str] ) -> Dict: '''simple docstring''' _UpperCamelCase = 20 _UpperCamelCase = model_class_name(lowerCAmelCase__ ) _UpperCamelCase = model.encode(inputs_dict['''input_ids'''] ) _UpperCamelCase , _UpperCamelCase = ( inputs_dict['''decoder_input_ids'''], inputs_dict['''decoder_attention_mask'''], ) _UpperCamelCase = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='''i4''' ) _UpperCamelCase = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _UpperCamelCase = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , decoder_position_ids=lowerCAmelCase__ , ) _UpperCamelCase = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='''i4''' ) _UpperCamelCase = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase__ , ) _UpperCamelCase = model.decode(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) def snake_case__ ( self : Any , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[int] ) -> Tuple: '''simple docstring''' _UpperCamelCase = 20 _UpperCamelCase = model_class_name(lowerCAmelCase__ ) _UpperCamelCase = model.encode(inputs_dict['''input_ids'''] ) _UpperCamelCase , _UpperCamelCase = ( inputs_dict['''decoder_input_ids'''], inputs_dict['''decoder_attention_mask'''], ) _UpperCamelCase = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _UpperCamelCase = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _UpperCamelCase = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ , decoder_position_ids=lowerCAmelCase__ , ) _UpperCamelCase = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='''i4''' ) _UpperCamelCase = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase__ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase__ , decoder_position_ids=lowerCAmelCase__ , ) _UpperCamelCase = model.decode(lowerCAmelCase__ , lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ ) _UpperCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=f"""Max diff is {diff}""" ) @require_flax class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" _snake_case : List[Any] = 9_9 def snake_case__ ( self : Tuple ) -> Dict: '''simple docstring''' _UpperCamelCase = np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) _UpperCamelCase = input_ids.shape[0] _UpperCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def snake_case__ ( self : int ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = self._get_config_and_data() _UpperCamelCase = FlaxBlenderbotSmallForConditionalGeneration(lowerCAmelCase__ ) _UpperCamelCase = lm_model(input_ids=lowerCAmelCase__ ) _UpperCamelCase = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs['''logits'''].shape , lowerCAmelCase__ ) def snake_case__ ( self : str ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) _UpperCamelCase = FlaxBlenderbotSmallForConditionalGeneration(lowerCAmelCase__ ) _UpperCamelCase = np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) _UpperCamelCase = np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) _UpperCamelCase = lm_model(input_ids=lowerCAmelCase__ , decoder_input_ids=lowerCAmelCase__ ) _UpperCamelCase = (*summary.shape, config.vocab_size) self.assertEqual(outputs['''logits'''].shape , lowerCAmelCase__ ) def snake_case__ ( self : Optional[int] ) -> Dict: '''simple docstring''' _UpperCamelCase = np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) _UpperCamelCase = shift_tokens_right(lowerCAmelCase__ , 1 , 2 ) _UpperCamelCase = np.equal(lowerCAmelCase__ , 1 ).astype(np.floataa ).sum() _UpperCamelCase = np.equal(lowerCAmelCase__ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCAmelCase__ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class __lowerCAmelCase ( __magic_name__ , unittest.TestCase , __magic_name__ ): """simple docstring""" _snake_case : List[Any] = True _snake_case : str = ( ( FlaxBlenderbotSmallModel, FlaxBlenderbotSmallForConditionalGeneration, ) if is_flax_available() else () ) _snake_case : Union[str, Any] = (FlaxBlenderbotSmallForConditionalGeneration,) if is_flax_available() else () def snake_case__ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' _UpperCamelCase = FlaxBlenderbotSmallModelTester(self ) def snake_case__ ( self : Optional[Any] ) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def snake_case__ ( self : Optional[int] ) -> Dict: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def snake_case__ ( self : int ) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _UpperCamelCase = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCamelCase = model_class(lowerCAmelCase__ ) @jax.jit def encode_jitted(lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : int=None , **lowerCAmelCase__ : Any ): return model.encode(input_ids=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ) with self.subTest('''JIT Enabled''' ): _UpperCamelCase = encode_jitted(**lowerCAmelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _UpperCamelCase = encode_jitted(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) ) for jitted_output, output in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self : Any ) -> Any: '''simple docstring''' _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _UpperCamelCase = model_class(lowerCAmelCase__ ) _UpperCamelCase = model.encode(inputs_dict['''input_ids'''] , inputs_dict['''attention_mask'''] ) _UpperCamelCase = { '''decoder_input_ids''': inputs_dict['''decoder_input_ids'''], '''decoder_attention_mask''': inputs_dict['''decoder_attention_mask'''], '''encoder_outputs''': encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase__ : str , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Dict ): return model.decode( decoder_input_ids=lowerCAmelCase__ , decoder_attention_mask=lowerCAmelCase__ , encoder_outputs=lowerCAmelCase__ , ) with self.subTest('''JIT Enabled''' ): _UpperCamelCase = decode_jitted(**lowerCAmelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _UpperCamelCase = decode_jitted(**lowerCAmelCase__ ).to_tuple() self.assertEqual(len(lowerCAmelCase__ ) , len(lowerCAmelCase__ ) ) for jitted_output, output in zip(lowerCAmelCase__ , lowerCAmelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def snake_case__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCamelCase = model_class_name.from_pretrained('''facebook/blenderbot_small-90M''' ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _UpperCamelCase = np.ones((1, 1) ) * model.config.eos_token_id _UpperCamelCase = model(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ )
98
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore SCREAMING_SNAKE_CASE : int = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" SCREAMING_SNAKE_CASE : Dict = [file for file in filepaths if file != file.lower()] if upper_files: print(F"{len(upper_files)} files contain uppercase characters:") print("\n".join(upper_files) + "\n") SCREAMING_SNAKE_CASE : List[Any] = [file for file in filepaths if " " in file] if space_files: print(F"{len(space_files)} files contain space characters:") print("\n".join(space_files) + "\n") SCREAMING_SNAKE_CASE : Any = [file for file in filepaths if "-" in file] if hyphen_files: print(F"{len(hyphen_files)} files contain hyphen characters:") print("\n".join(hyphen_files) + "\n") SCREAMING_SNAKE_CASE : str = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"{len(nodir_files)} files are not in a directory:") print("\n".join(nodir_files) + "\n") SCREAMING_SNAKE_CASE : Tuple = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
89
0
from string import ascii_uppercase SCREAMING_SNAKE_CASE = {char: i for i, char in enumerate(ascii_uppercase)} SCREAMING_SNAKE_CASE = dict(enumerate(ascii_uppercase)) def a (lowerCAmelCase__ , lowerCAmelCase__ ): __a = len(lowerCAmelCase__ ) __a = 0 while True: if x == i: __a = 0 if len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ): break key += key[i] i += 1 return key def a (lowerCAmelCase__ , lowerCAmelCase__ ): __a = """""" __a = 0 for letter in message: if letter == " ": cipher_text += " " else: __a = (dicta[letter] - dicta[key_new[i]]) % 26 i += 1 cipher_text += dicta[x] return cipher_text def a (lowerCAmelCase__ , lowerCAmelCase__ ): __a = """""" __a = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: __a = (dicta[letter] + dicta[key_new[i]] + 26) % 26 i += 1 or_txt += dicta[x] return or_txt def a (): __a = """THE GERMAN ATTACK""" __a = """SECRET""" __a = generate_key(lowerCAmelCase__ , lowerCAmelCase__ ) __a = cipher_text(lowerCAmelCase__ , lowerCAmelCase__ ) print(f'''Encrypted Text = {s}''' ) print(f'''Original Text = {original_text(lowerCAmelCase__ , lowerCAmelCase__ )}''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
99
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Any: _lowercase : str = 10 _lowercase : List[str] = datasets.Features( { 'tokens': datasets.Sequence(datasets.Value('string' ) ), 'labels': datasets.Sequence(datasets.ClassLabel(names=['negative', 'positive'] ) ), 'answers': datasets.Sequence( { 'text': datasets.Value('string' ), 'answer_start': datasets.Value('int32' ), } ), 'id': datasets.Value('int64' ), } ) _lowercase : Union[str, Any] = datasets.Dataset.from_dict( { 'tokens': [['foo'] * 5] * n, 'labels': [[1] * 5] * n, 'answers': [{'answer_start': [97], 'text': ['1976']}] * 10, 'id': list(range(lowerCamelCase_ ) ), } , features=lowerCamelCase_ , ) return dataset @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : int = str(tmp_path_factory.mktemp('data' ) / 'file.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return filename # FILE_CONTENT + files SCREAMING_SNAKE_CASE : str = "\\n Text data.\n Second line of data." @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'file.txt' _lowercase : List[str] = FILE_CONTENT with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Tuple: import bza _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.bz2' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'file.txt.gz' ) _lowercase : Optional[int] = bytes(lowerCamelCase_ , 'utf-8' ) with gzip.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: if datasets.config.LZ4_AVAILABLE: import lza.frame _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'file.txt.lz4' _lowercase : Optional[Any] = bytes(lowerCamelCase_ , 'utf-8' ) with lza.frame.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: if datasets.config.PY7ZR_AVAILABLE: import pyazr _lowercase : int = tmp_path_factory.mktemp('data' ) / 'file.txt.7z' with pyazr.SevenZipFile(lowerCamelCase_ , 'w' ) as archive: archive.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: import tarfile _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: import lzma _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.xz' _lowercase : int = bytes(lowerCamelCase_ , 'utf-8' ) with lzma.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> str: import zipfile _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'file.txt.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'file.txt.zst' _lowercase : Dict = bytes(lowerCamelCase_ , 'utf-8' ) with zstd.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'file.xml' _lowercase : Optional[Any] = textwrap.dedent( '\\n <?xml version="1.0" encoding="UTF-8" ?>\n <tmx version="1.4">\n <header segtype="sentence" srclang="ca" />\n <body>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 1</seg></tuv>\n <tuv xml:lang="en"><seg>Content 1</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 2</seg></tuv>\n <tuv xml:lang="en"><seg>Content 2</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 3</seg></tuv>\n <tuv xml:lang="en"><seg>Content 3</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 4</seg></tuv>\n <tuv xml:lang="en"><seg>Content 4</seg></tuv>\n </tu>\n <tu>\n <tuv xml:lang="ca"><seg>Contingut 5</seg></tuv>\n <tuv xml:lang="en"><seg>Content 5</seg></tuv>\n </tu>\n </body>\n </tmx>' ) with open(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ ) return filename SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] SCREAMING_SNAKE_CASE : Dict = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] SCREAMING_SNAKE_CASE : Optional[Any] = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } SCREAMING_SNAKE_CASE : Tuple = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] SCREAMING_SNAKE_CASE : Any = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> List[str]: return DATA_DICT_OF_LISTS @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Optional[int] = datasets.Dataset.from_dict(lowerCamelCase_ ) _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.arrow' ) dataset.map(cache_file_name=lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> str: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.sqlite' ) with contextlib.closing(sqlitea.connect(lowerCamelCase_ ) ) as con: _lowercase : Union[str, Any] = con.cursor() cur.execute('CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)' ) for item in DATA: cur.execute('INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)' , tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : List[str] = str(tmp_path_factory.mktemp('data' ) / 'dataset.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : Tuple = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset2.csv' ) with open(lowerCamelCase_ , 'w' , newline='' ) as f: _lowercase : str = csv.DictWriter(lowerCamelCase_ , fieldnames=['col_1', 'col_2', 'col_3'] ) writer.writeheader() for item in DATA: writer.writerow(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: import bza _lowercase : int = tmp_path_factory.mktemp('data' ) / 'dataset.csv.bz2' with open(lowerCamelCase_ , 'rb' ) as f: _lowercase : int = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(lowerCamelCase_ , 'wb' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(csv_path.replace('.csv' , '.CSV' ) ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(csva_path.replace('.csv' , '.CSV' ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Union[str, Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.csv.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.parquet' ) _lowercase : Optional[Any] = pa.schema( { 'col_1': pa.string(), 'col_2': pa.intaa(), 'col_3': pa.floataa(), } ) with open(lowerCamelCase_ , 'wb' ) as f: _lowercase : List[str] = pq.ParquetWriter(lowerCamelCase_ , schema=lowerCamelCase_ ) _lowercase : Any = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(lowerCamelCase_ ) )] for k in DATA[0]} , schema=lowerCamelCase_ ) writer.write_table(lowerCamelCase_ ) writer.close() return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : List[Any] = {'data': DATA} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.json' ) _lowercase : Optional[Any] = {'data': DATA_DICT_OF_LISTS} with open(lowerCamelCase_ , 'w' ) as f: json.dump(lowerCamelCase_ , lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: _lowercase : Dict = str(tmp_path_factory.mktemp('data' ) / 'dataset2.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : Optional[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset_312.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_312: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[Any]: _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset-str.jsonl' ) with open(lowerCamelCase_ , 'w' ) as f: for item in DATA_STR: f.write(json.dumps(lowerCamelCase_ ) + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Optional[Any]: import gzip _lowercase : Tuple = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Dict: import gzip _lowercase : Optional[int] = str(tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.gz' ) with open(lowerCamelCase_ , 'rb' ) as orig_file: with gzip.open(lowerCamelCase_ , 'wb' ) as zipped_file: zipped_file.writelines(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.jsonl.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : Tuple = tmp_path_factory.mktemp('data' ) / 'dataset.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.add(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : str = tmp_path_factory.mktemp('data' ) / 'dataset_nested.jsonl.tar' with tarfile.TarFile(lowerCamelCase_ , 'w' ) as f: f.add(lowerCamelCase_ , arcname=os.path.join('nested' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : Optional[int] = ['0', '1', '2', '3'] _lowercase : str = str(tmp_path_factory.mktemp('data' ) / 'dataset.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Union[str, Any]: _lowercase : str = ['0', '1', '2', '3'] _lowercase : List[Any] = str(tmp_path_factory.mktemp('data' ) / 'dataset2.txt' ) with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> List[str]: _lowercase : List[Any] = ['0', '1', '2', '3'] _lowercase : Optional[int] = tmp_path_factory.mktemp('data' ) / 'dataset.abc' with open(lowerCamelCase_ , 'w' ) as f: for item in data: f.write(item + '\n' ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Optional[Any] = tmp_path_factory.mktemp('data' ) / 'dataset.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: _lowercase : List[Any] = tmp_path_factory.mktemp('data' ) / 'dataset_with_dir.text.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) f.write(lowerCamelCase_ , arcname=os.path.join('main_dir' , os.path.basename(lowerCamelCase_ ) ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: _lowercase : Any = tmp_path_factory.mktemp('data' ) / 'dataset.ext.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported.ext' ) ) f.write(lowerCamelCase_ , arcname=os.path.basename('unsupported_2.ext' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = '\n'.join(['First', 'Second\u2029with Unicode new line', 'Third'] ) _lowercase : Any = str(tmp_path_factory.mktemp('data' ) / 'dataset_with_unicode_new_lines.txt' ) with open(lowerCamelCase_ , 'w' , encoding='utf-8' ) as f: f.write(lowerCamelCase_ ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> Dict: return os.path.join('tests' , 'features' , 'data' , 'test_image_rgb.jpg' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( ) -> int: return os.path.join('tests' , 'features' , 'data' , 'test_audio_44100.wav' ) @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Any: _lowercase : Dict = tmp_path_factory.mktemp('data' ) / 'dataset.img.zip' with zipfile.ZipFile(lowerCamelCase_ , 'w' ) as f: f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ) ) f.write(lowerCamelCase_ , arcname=os.path.basename(lowerCamelCase_ ).replace('.jpg' , '2.jpg' ) ) return path @pytest.fixture(scope='session' ) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[Any]: _lowercase : str = tmp_path_factory.mktemp('data_dir' ) (data_dir / "subdir").mkdir() with open(data_dir / 'subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / 'subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden file with open(data_dir / 'subdir' / '.test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / '.subdir' / 'train.txt' , 'w' ) as f: f.write('foo\n' * 10 ) with open(data_dir / '.subdir' / 'test.txt' , 'w' ) as f: f.write('bar\n' * 10 ) return data_dir
89
0
import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() _A : List[str] = logging.get_logger(__name__) def __snake_case ( lowerCAmelCase_ ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = SwinConfig.from_pretrained( '''microsoft/swin-tiny-patch4-window7-224''' , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) SCREAMING_SNAKE_CASE__ = MaskFormerConfig(backbone_config=lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = '''huggingface/label-files''' if "ade20k-full" in model_name: # this should be ok SCREAMING_SNAKE_CASE__ = 8_4_7 SCREAMING_SNAKE_CASE__ = '''maskformer-ade20k-full-id2label.json''' elif "ade" in model_name: # this should be ok SCREAMING_SNAKE_CASE__ = 1_5_0 SCREAMING_SNAKE_CASE__ = '''ade20k-id2label.json''' elif "coco-stuff" in model_name: # this should be ok SCREAMING_SNAKE_CASE__ = 1_7_1 SCREAMING_SNAKE_CASE__ = '''maskformer-coco-stuff-id2label.json''' elif "coco" in model_name: # TODO SCREAMING_SNAKE_CASE__ = 1_3_3 SCREAMING_SNAKE_CASE__ = '''coco-panoptic-id2label.json''' elif "cityscapes" in model_name: # this should be ok SCREAMING_SNAKE_CASE__ = 1_9 SCREAMING_SNAKE_CASE__ = '''cityscapes-id2label.json''' elif "vistas" in model_name: # this should be ok SCREAMING_SNAKE_CASE__ = 6_5 SCREAMING_SNAKE_CASE__ = '''mapillary-vistas-id2label.json''' SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type='''dataset''' ) , '''r''' ) ) SCREAMING_SNAKE_CASE__ = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} return config def __snake_case ( lowerCAmelCase_ ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = [] # stem # fmt: off rename_keys.append(('''backbone.patch_embed.proj.weight''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.proj.bias''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''model.pixel_level_module.encoder.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''model.pixel_level_module.encoder.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm1.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm1.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.relative_position_index''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.proj.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.attn.proj.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm2.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.norm2.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc1.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc1.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc2.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((f'''backbone.layers.{i}.blocks.{j}.mlp.fc2.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((f'''backbone.layers.{i}.downsample.reduction.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((f'''backbone.layers.{i}.downsample.norm.weight''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((f'''backbone.layers.{i}.downsample.norm.bias''', f'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((f'''backbone.norm{i}.weight''', f'''model.pixel_level_module.encoder.hidden_states_norms.{i}.weight''') ) rename_keys.append((f'''backbone.norm{i}.bias''', f'''model.pixel_level_module.encoder.hidden_states_norms.{i}.bias''') ) # FPN rename_keys.append(('''sem_seg_head.layer_4.weight''', '''model.pixel_level_module.decoder.fpn.stem.0.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.weight''', '''model.pixel_level_module.decoder.fpn.stem.1.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.bias''', '''model.pixel_level_module.decoder.fpn.stem.1.bias''') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((f'''sem_seg_head.adapter_{source_index}.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight''') ) rename_keys.append((f'''sem_seg_head.adapter_{source_index}.norm.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight''') ) rename_keys.append((f'''sem_seg_head.adapter_{source_index}.norm.bias''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias''') ) rename_keys.append((f'''sem_seg_head.layer_{source_index}.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight''') ) rename_keys.append((f'''sem_seg_head.layer_{source_index}.norm.weight''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight''') ) rename_keys.append((f'''sem_seg_head.layer_{source_index}.norm.bias''', f'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias''') ) rename_keys.append(('''sem_seg_head.mask_features.weight''', '''model.pixel_level_module.decoder.mask_projection.weight''') ) rename_keys.append(('''sem_seg_head.mask_features.bias''', '''model.pixel_level_module.decoder.mask_projection.bias''') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight''', f'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias''', f'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias''') ) # cross-attention out projection rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias''') ) # MLP 1 rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight''', f'''model.transformer_module.decoder.layers.{idx}.fc1.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias''', f'''model.transformer_module.decoder.layers.{idx}.fc1.bias''') ) # MLP 2 rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight''', f'''model.transformer_module.decoder.layers.{idx}.fc2.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias''', f'''model.transformer_module.decoder.layers.{idx}.fc2.bias''') ) # layernorm 1 (self-attention layernorm) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight''', f'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias''', f'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias''') ) # layernorm 2 (cross-attention layernorm) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias''', f'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias''') ) # layernorm 3 (final layernorm) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight''', f'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias''', f'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.weight''', '''model.transformer_module.decoder.layernorm.weight''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.bias''', '''model.transformer_module.decoder.layernorm.bias''') ) # heads on top rename_keys.append(('''sem_seg_head.predictor.query_embed.weight''', '''model.transformer_module.queries_embedder.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.weight''', '''model.transformer_module.input_projection.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.bias''', '''model.transformer_module.input_projection.bias''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.weight''', '''class_predictor.weight''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.bias''', '''class_predictor.bias''') ) for i in range(3 ): rename_keys.append((f'''sem_seg_head.predictor.mask_embed.layers.{i}.weight''', f'''mask_embedder.{i}.0.weight''') ) rename_keys.append((f'''sem_seg_head.predictor.mask_embed.layers.{i}.bias''', f'''mask_embedder.{i}.0.bias''') ) # fmt: on return rename_keys def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> str: SCREAMING_SNAKE_CASE__ = dct.pop(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = val def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Any: SCREAMING_SNAKE_CASE__ = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): SCREAMING_SNAKE_CASE__ = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) SCREAMING_SNAKE_CASE__ = state_dict.pop(f'''backbone.layers.{i}.blocks.{j}.attn.qkv.weight''' ) SCREAMING_SNAKE_CASE__ = state_dict.pop(f'''backbone.layers.{i}.blocks.{j}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE__ = in_proj_weight[:dim, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[: dim] SCREAMING_SNAKE_CASE__ = in_proj_weight[ dim : dim * 2, : ] SCREAMING_SNAKE_CASE__ = in_proj_bias[ dim : dim * 2 ] SCREAMING_SNAKE_CASE__ = in_proj_weight[ -dim :, : ] SCREAMING_SNAKE_CASE__ = in_proj_bias[-dim :] # fmt: on def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Tuple: # fmt: off SCREAMING_SNAKE_CASE__ = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) SCREAMING_SNAKE_CASE__ = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight''' ) SCREAMING_SNAKE_CASE__ = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE__ = in_proj_weight[: hidden_size, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[:config.hidden_size] SCREAMING_SNAKE_CASE__ = in_proj_weight[hidden_size : hidden_size * 2, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[hidden_size : hidden_size * 2] SCREAMING_SNAKE_CASE__ = in_proj_weight[-hidden_size :, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) SCREAMING_SNAKE_CASE__ = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight''' ) SCREAMING_SNAKE_CASE__ = state_dict.pop(f'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE__ = in_proj_weight[: hidden_size, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[:config.hidden_size] SCREAMING_SNAKE_CASE__ = in_proj_weight[hidden_size : hidden_size * 2, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[hidden_size : hidden_size * 2] SCREAMING_SNAKE_CASE__ = in_proj_weight[-hidden_size :, :] SCREAMING_SNAKE_CASE__ = in_proj_bias[-hidden_size :] # fmt: on def __snake_case ( ) -> torch.Tensor: SCREAMING_SNAKE_CASE__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' SCREAMING_SNAKE_CASE__ = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = False ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = get_maskformer_config(lowerCAmelCase_ ) # load original state_dict with open(lowerCAmelCase_ , '''rb''' ) as f: SCREAMING_SNAKE_CASE__ = pickle.load(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = data['''model'''] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys SCREAMING_SNAKE_CASE__ = create_rename_keys(lowerCAmelCase_ ) for src, dest in rename_keys: rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) read_in_swin_q_k_v(lowerCAmelCase_ , config.backbone_config ) read_in_decoder_q_k_v(lowerCAmelCase_ , lowerCAmelCase_ ) # update to torch tensors for key, value in state_dict.items(): SCREAMING_SNAKE_CASE__ = torch.from_numpy(lowerCAmelCase_ ) # load 🤗 model SCREAMING_SNAKE_CASE__ = MaskFormerForInstanceSegmentation(lowerCAmelCase_ ) model.eval() for name, param in model.named_parameters(): print(lowerCAmelCase_ , param.shape ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = model.load_state_dict(lowerCAmelCase_ , strict=lowerCAmelCase_ ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(lowerCAmelCase_ ) == 0, f'''Unexpected keys: {unexpected_keys}''' # verify results SCREAMING_SNAKE_CASE__ = prepare_img() if "vistas" in model_name: SCREAMING_SNAKE_CASE__ = 6_5 elif "cityscapes" in model_name: SCREAMING_SNAKE_CASE__ = 6_5_5_3_5 else: SCREAMING_SNAKE_CASE__ = 2_5_5 SCREAMING_SNAKE_CASE__ = True if '''ade''' in model_name else False SCREAMING_SNAKE_CASE__ = MaskFormerImageProcessor(ignore_index=lowerCAmelCase_ , reduce_labels=lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = image_processor(lowerCAmelCase_ , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE__ = model(**lowerCAmelCase_ ) print('''Logits:''' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": SCREAMING_SNAKE_CASE__ = torch.tensor( [[3.63_53, -4.47_70, -2.60_65], [0.50_81, -4.23_94, -3.53_43], [2.19_09, -5.03_53, -1.93_23]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase_ , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f'''Saving model and image processor to {pytorch_dump_folder_path}''' ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) image_processor.save_pretrained(lowerCAmelCase_ ) if push_to_hub: print('''Pushing model and image processor to the hub...''' ) model.push_to_hub(f'''nielsr/{model_name}''' ) image_processor.push_to_hub(f'''nielsr/{model_name}''' ) if __name__ == "__main__": _A : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""maskformer-swin-tiny-ade""", type=str, help=("""Name of the MaskFormer model you'd like to convert""",), ) parser.add_argument( """--checkpoint_path""", default="""/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl""", type=str, help="""Path to the original state dict (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _A : str = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
100
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE : str = { "configuration_gpt_bigcode": ["GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTBigCodeConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : int = [ "GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTBigCodeForSequenceClassification", "GPTBigCodeForTokenClassification", "GPTBigCodeForCausalLM", "GPTBigCodeModel", "GPTBigCodePreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : Dict = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 1_8, 2] SCREAMING_SNAKE_CASE_ : Union[str, Any] = True if 'large' in model_name or 'huge' in model_name else False SCREAMING_SNAKE_CASE_ : List[str] = True if 'large' in model_name or 'huge' in model_name else False SCREAMING_SNAKE_CASE_ : Tuple = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: SCREAMING_SNAKE_CASE_ : Optional[int] = [3, 3, 3, 3] SCREAMING_SNAKE_CASE_ : Optional[int] = [5, 5, 5, 5] elif "fl4" in model_name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = [4, 4, 4, 4] SCREAMING_SNAKE_CASE_ : Tuple = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: SCREAMING_SNAKE_CASE_ : str = [3, 3, 3, 3] if "lrf" in model_name: SCREAMING_SNAKE_CASE_ : Any = [3, 3, 3, 3] else: SCREAMING_SNAKE_CASE_ : Optional[int] = [2, 2, 2, 2] if "tiny" in model_name: SCREAMING_SNAKE_CASE_ : Any = 9_6 elif "small" in model_name: SCREAMING_SNAKE_CASE_ : str = 9_6 elif "base" in model_name: SCREAMING_SNAKE_CASE_ : Any = 1_2_8 elif "large" in model_name: SCREAMING_SNAKE_CASE_ : List[Any] = 1_9_2 elif "xlarge" in model_name: SCREAMING_SNAKE_CASE_ : Optional[int] = 2_5_6 elif "huge" in model_name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = 3_5_2 # set label information SCREAMING_SNAKE_CASE_ : Union[str, Any] = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: SCREAMING_SNAKE_CASE_ : Optional[Any] = 'imagenet-22k-id2label.json' else: SCREAMING_SNAKE_CASE_ : Optional[int] = 'imagenet-1k-id2label.json' SCREAMING_SNAKE_CASE_ : Any = json.load(open(hf_hub_download(A__, A__, repo_type='dataset' ), 'r' ) ) SCREAMING_SNAKE_CASE_ : str = {int(A__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ : Optional[int] = {v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE_ : int = FocalNetConfig( embed_dim=A__, depths=A__, focal_levels=A__, focal_windows=A__, use_conv_embed=A__, idalabel=A__, labelaid=A__, use_post_layernorm=A__, use_layerscale=A__, ) return config def a__ ( A__ ): if "patch_embed.proj" in name: SCREAMING_SNAKE_CASE_ : str = name.replace('patch_embed.proj', 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: SCREAMING_SNAKE_CASE_ : int = name.replace('patch_embed.norm', 'embeddings.norm' ) if "layers" in name: SCREAMING_SNAKE_CASE_ : Any = 'encoder.' + name if "encoder.layers" in name: SCREAMING_SNAKE_CASE_ : Optional[int] = name.replace('encoder.layers', 'encoder.stages' ) if "downsample.proj" in name: SCREAMING_SNAKE_CASE_ : str = name.replace('downsample.proj', 'downsample.projection' ) if "blocks" in name: SCREAMING_SNAKE_CASE_ : Any = name.replace('blocks', 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: SCREAMING_SNAKE_CASE_ : Optional[int] = name.replace('modulation.f', 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: SCREAMING_SNAKE_CASE_ : Dict = name.replace('modulation.h', 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: SCREAMING_SNAKE_CASE_ : Optional[Any] = name.replace('modulation.proj', 'modulation.projection_out' ) if name == "norm.weight": SCREAMING_SNAKE_CASE_ : int = 'layernorm.weight' if name == "norm.bias": SCREAMING_SNAKE_CASE_ : List[Any] = 'layernorm.bias' if "head" in name: SCREAMING_SNAKE_CASE_ : List[Any] = name.replace('head', 'classifier' ) else: SCREAMING_SNAKE_CASE_ : List[str] = 'focalnet.' + name return name def a__ ( A__, A__, A__=False ): # fmt: off SCREAMING_SNAKE_CASE_ : List[Any] = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on SCREAMING_SNAKE_CASE_ : Optional[int] = model_name_to_url[model_name] print('Checkpoint URL: ', A__ ) SCREAMING_SNAKE_CASE_ : str = torch.hub.load_state_dict_from_url(A__, map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): SCREAMING_SNAKE_CASE_ : Dict = state_dict.pop(A__ ) SCREAMING_SNAKE_CASE_ : Dict = val SCREAMING_SNAKE_CASE_ : List[Any] = get_focalnet_config(A__ ) SCREAMING_SNAKE_CASE_ : int = FocalNetForImageClassification(A__ ) model.eval() # load state dict model.load_state_dict(A__ ) # verify conversion SCREAMING_SNAKE_CASE_ : Optional[int] = 'http://images.cocodataset.org/val2017/000000039769.jpg' SCREAMING_SNAKE_CASE_ : List[Any] = BitImageProcessor( do_resize=A__, size={'shortest_edge': 2_5_6}, resample=PILImageResampling.BILINEAR, do_center_crop=A__, crop_size=2_2_4, do_normalize=A__, image_mean=A__, image_std=A__, ) SCREAMING_SNAKE_CASE_ : List[Any] = Image.open(requests.get(A__, stream=A__ ).raw ) SCREAMING_SNAKE_CASE_ : str = processor(images=A__, return_tensors='pt' ) SCREAMING_SNAKE_CASE_ : int = transforms.Compose( [ transforms.Resize(2_5_6 ), transforms.CenterCrop(2_2_4 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06], std=[0.2_29, 0.2_24, 0.2_25] ), ] ) SCREAMING_SNAKE_CASE_ : Optional[Any] = image_transforms(A__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values, A__, atol=1E-4 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = model(**A__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = outputs.logits.argmax(-1 ).item() print('Predicted class:', model.config.idalabel[predicted_class_idx] ) print('First values of logits:', outputs.logits[0, :3] ) if model_name == "focalnet-tiny": SCREAMING_SNAKE_CASE_ : Dict = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": SCREAMING_SNAKE_CASE_ : Any = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": SCREAMING_SNAKE_CASE_ : List[Any] = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": SCREAMING_SNAKE_CASE_ : Optional[int] = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": SCREAMING_SNAKE_CASE_ : Optional[int] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3], A__, atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(A__ ) processor.save_pretrained(A__ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": lowerCAmelCase__ : List[Any] =argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='focalnet-tiny', type=str, help='Name of the FocalNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub.', ) lowerCAmelCase__ : Optional[int] =parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
101
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=7, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=False, lowerCamelCase=2, lowerCamelCase=99, lowerCamelCase=0, lowerCamelCase=32, lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=0.1, lowerCamelCase=0.1, lowerCamelCase=5_12, lowerCamelCase=2, lowerCamelCase=0.0_2, lowerCamelCase=2, lowerCamelCase=4, lowerCamelCase="last", lowerCamelCase=True, lowerCamelCase=None, lowerCamelCase=0, ) -> str: """simple docstring""" _lowercase : Union[str, Any] = parent _lowercase : Optional[Any] = batch_size _lowercase : List[str] = seq_length _lowercase : int = is_training _lowercase : List[str] = use_input_lengths _lowercase : int = use_token_type_ids _lowercase : Any = use_labels _lowercase : Union[str, Any] = gelu_activation _lowercase : List[str] = sinusoidal_embeddings _lowercase : str = causal _lowercase : Optional[int] = asm _lowercase : Union[str, Any] = n_langs _lowercase : List[Any] = vocab_size _lowercase : Any = n_special _lowercase : Any = hidden_size _lowercase : str = num_hidden_layers _lowercase : Union[str, Any] = num_attention_heads _lowercase : Tuple = hidden_dropout_prob _lowercase : Optional[int] = attention_probs_dropout_prob _lowercase : Union[str, Any] = max_position_embeddings _lowercase : List[str] = type_sequence_label_size _lowercase : Any = initializer_range _lowercase : int = num_labels _lowercase : Optional[int] = num_choices _lowercase : Optional[Any] = summary_type _lowercase : Optional[Any] = use_proj _lowercase : int = scope _lowercase : List[Any] = bos_token_id def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Tuple = ids_tensor([self.batch_size, self.seq_length], self.vocab_size) _lowercase : List[str] = random_attention_mask([self.batch_size, self.seq_length]) _lowercase : int = None if self.use_input_lengths: _lowercase : Dict = ( ids_tensor([self.batch_size], vocab_size=2) + self.seq_length - 2 ) # small variation of seq_length _lowercase : Tuple = None if self.use_token_type_ids: _lowercase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length], self.n_langs) _lowercase : Tuple = None _lowercase : int = None _lowercase : int = None if self.use_labels: _lowercase : str = ids_tensor([self.batch_size], self.type_sequence_label_size) _lowercase : str = ids_tensor([self.batch_size, self.seq_length], self.num_labels) _lowercase : Dict = ids_tensor([self.batch_size], 2).float() _lowercase : Tuple = ids_tensor([self.batch_size], self.num_choices) _lowercase : Dict = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return XLMConfig( vocab_size=self.vocab_size, n_special=self.n_special, emb_dim=self.hidden_size, n_layers=self.num_hidden_layers, n_heads=self.num_attention_heads, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, gelu_activation=self.gelu_activation, sinusoidal_embeddings=self.sinusoidal_embeddings, asm=self.asm, causal=self.causal, n_langs=self.n_langs, max_position_embeddings=self.max_position_embeddings, initializer_range=self.initializer_range, summary_type=self.summary_type, use_proj=self.use_proj, num_labels=self.num_labels, bos_token_id=self.bos_token_id, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Tuple: """simple docstring""" _lowercase : List[Any] = XLMModel(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : str = model(lowerCamelCase, lengths=lowerCamelCase, langs=lowerCamelCase) _lowercase : int = model(lowerCamelCase, langs=lowerCamelCase) _lowercase : Any = model(lowerCamelCase) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[Any]: """simple docstring""" _lowercase : Dict = XLMWithLMHeadModel(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> str: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnsweringSimple(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Dict = model(lowerCamelCase) _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) _lowercase : Any = outputs self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMForQuestionAnswering(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = model(lowerCamelCase) _lowercase : List[Any] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, p_mask=lowerCamelCase, ) _lowercase : List[str] = model( lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, cls_index=lowerCamelCase, is_impossible=lowerCamelCase, ) ((_lowercase) , ) : Optional[Any] = result_with_labels.to_tuple() _lowercase : List[str] = model(lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase) ((_lowercase) , ) : Any = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape, ()) self.parent.assertEqual(result.start_top_log_probs.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual(result.start_top_index.shape, (self.batch_size, model.config.start_n_top)) self.parent.assertEqual( result.end_top_log_probs.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual( result.end_top_index.shape, (self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual(result.cls_logits.shape, (self.batch_size,)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> int: """simple docstring""" _lowercase : Optional[Any] = XLMForSequenceClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[int] = model(lowerCamelCase) _lowercase : Optional[int] = model(lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.loss.shape, ()) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.type_sequence_label_size)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> List[str]: """simple docstring""" _lowercase : Any = self.num_labels _lowercase : str = XLMForTokenClassification(lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : int = model(lowerCamelCase, attention_mask=lowerCamelCase, labels=lowerCamelCase) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.num_labels)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, ) -> Dict: """simple docstring""" _lowercase : Optional[Any] = self.num_choices _lowercase : Optional[int] = XLMForMultipleChoice(config=lowerCamelCase) model.to(lowerCamelCase) model.eval() _lowercase : Optional[Any] = input_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : int = token_type_ids.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : Optional[Any] = input_mask.unsqueeze(1).expand(-1, self.num_choices, -1).contiguous() _lowercase : List[str] = model( lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_choices)) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Dict = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Optional[Any] = config_and_inputs _lowercase : List[str] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return config, inputs_dict @require_torch class _lowerCamelCase( _a, _a, _a, unittest.TestCase ): lowercase_ : Any = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) lowercase_ : Optional[int] = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable lowercase_ : Union[str, Any] = ( { """feature-extraction""": XLMModel, """fill-mask""": XLMWithLMHeadModel, """question-answering""": XLMForQuestionAnsweringSimple, """text-classification""": XLMForSequenceClassification, """text-generation""": XLMWithLMHeadModel, """token-classification""": XLMForTokenClassification, """zero-shot""": XLMForSequenceClassification, } if is_torch_available() else {} ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> Optional[int]: """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast') ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase=False) -> Optional[int]: """simple docstring""" _lowercase : Any = super()._prepare_for_class(lowerCamelCase, lowerCamelCase, return_labels=lowerCamelCase) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": _lowercase : Any = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) _lowercase : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase) return inputs_dict def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Union[str, Any] = XLMModelTester(self) _lowercase : List[str] = ConfigTester(self, config_class=lowerCamelCase, emb_dim=37) def UpperCamelCase ( self) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*lowerCamelCase) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*lowerCamelCase) def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*lowerCamelCase) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*lowerCamelCase) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> int: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_attentions in attentions], [True] * len(lowerCamelCase)) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_attentions in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : Dict = min_length + idx + 1 _lowercase : int = min_length + idx + 1 _lowercase : Dict = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions], [expected_shape] * len(lowerCamelCase)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=False, lowerCamelCase=1) -> List[Any]: """simple docstring""" self.assertIsInstance(lowerCamelCase, lowerCamelCase) self.assertListEqual( [isinstance(lowerCamelCase, lowerCamelCase) for iter_hidden_states in hidden_states], [True] * len(lowerCamelCase), ) self.assertEqual(len(lowerCamelCase), (max_length - min_length) * num_beam_groups) for idx, iter_hidden_states in enumerate(lowerCamelCase): # adds PAD dummy token _lowercase : int = min_length + idx + 1 _lowercase : int = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states], [expected_shape] * len(lowerCamelCase), ) pass @slow def UpperCamelCase ( self) -> int: """simple docstring""" for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Dict = XLMModel.from_pretrained(lowerCamelCase) self.assertIsNotNone(lowerCamelCase) @require_torch class _lowerCamelCase( unittest.TestCase ): @slow def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Tuple = XLMWithLMHeadModel.from_pretrained('xlm-mlm-en-2048') model.to(lowerCamelCase) _lowercase : Union[str, Any] = torch.tensor([[14, 4_47]], dtype=torch.long, device=lowerCamelCase) # the president _lowercase : Any = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference _lowercase : str = model.generate(lowerCamelCase, do_sample=lowerCamelCase) self.assertListEqual(output_ids[0].cpu().numpy().tolist(), lowerCamelCase)
89
0
"""simple docstring""" import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel __magic_name__ : Any = { """gwf-440k""": { """url""": """https://model-server.zqevans2.workers.dev/gwf-440k.ckpt""", """sample_rate""": 4_8_0_0_0, """sample_size""": 6_5_5_3_6, }, """jmann-small-190k""": { """url""": """https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt""", """sample_rate""": 4_8_0_0_0, """sample_size""": 6_5_5_3_6, }, """jmann-large-580k""": { """url""": """https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt""", """sample_rate""": 4_8_0_0_0, """sample_size""": 1_3_1_0_7_2, }, """maestro-uncond-150k""": { """url""": """https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt""", """sample_rate""": 1_6_0_0_0, """sample_size""": 6_5_5_3_6, }, """unlocked-uncond-250k""": { """url""": """https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt""", """sample_rate""": 1_6_0_0_0, """sample_size""": 6_5_5_3_6, }, """honk-140k""": { """url""": """https://model-server.zqevans2.workers.dev/honk-140k.ckpt""", """sample_rate""": 1_6_0_0_0, """sample_size""": 6_5_5_3_6, }, } def UpperCamelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return torch.atana(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) / math.pi * 2 def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : Optional[Any] = torch.sin(t * math.pi / 2 ) ** 2 UpperCamelCase : Dict = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) class lowercase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" pass class lowercase__ ( nn.Module ): """simple docstring""" def __init__( self , _A ): '''simple docstring''' super().__init__() UpperCamelCase : List[str] = DiffusionAttnUnetaD(_A , n_attn_layers=4 ) UpperCamelCase : Tuple = deepcopy(self.diffusion ) UpperCamelCase : Dict = torch.quasirandom.SobolEngine(1 , scramble=_A ) def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : Optional[int] = MODELS_MAP[model_name]["""url"""] os.system(f"""wget {url} ./""" ) return f"""./{model_name}.ckpt""" __magic_name__ : str = { """1""": """resnets.0""", """2""": """attentions.0""", """3""": """resnets.1""", """4""": """attentions.1""", """5""": """resnets.2""", """6""": """attentions.2""", } __magic_name__ : Dict = { """8""": """resnets.0""", """9""": """attentions.0""", """10""": """resnets.1""", """11""": """attentions.1""", """12""": """resnets.2""", """13""": """attentions.2""", } __magic_name__ : Any = { """1""": """resnets.0""", """2""": """attentions.0""", """3""": """resnets.1""", """4""": """attentions.1""", """5""": """resnets.2""", """6""": """attentions.2""", """8""": """resnets.3""", """9""": """attentions.3""", """10""": """resnets.4""", """11""": """attentions.4""", """12""": """resnets.5""", """13""": """attentions.5""", } __magic_name__ : Dict = { """0""": """resnets.0""", """1""": """resnets.1""", """2""": """resnets.2""", """4""": """resnets.0""", """5""": """resnets.1""", """6""": """resnets.2""", } __magic_name__ : Any = { """skip""": """conv_skip""", """main.0""": """conv_1""", """main.1""": """group_norm_1""", """main.3""": """conv_2""", """main.4""": """group_norm_2""", } __magic_name__ : Any = { """norm""": """group_norm""", """qkv_proj""": ["""query""", """key""", """value"""], """out_proj""": ["""proj_attn"""], } def UpperCamelCase (SCREAMING_SNAKE_CASE ): if name.startswith("""skip""" ): return name.replace("""skip""" , RES_CONV_MAP["""skip"""] ) # name has to be of format main.{digit} if not name.startswith("""main.""" ): raise ValueError(f"""ResConvBlock error with {name}""" ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def UpperCamelCase (SCREAMING_SNAKE_CASE ): for key, value in ATTN_MAP.items(): if name.startswith(SCREAMING_SNAKE_CASE ) and not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return name.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif name.startswith(SCREAMING_SNAKE_CASE ): return [name.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for v in value] raise ValueError(f"""Attn error with {name}""" ) def UpperCamelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=13 ): UpperCamelCase : Optional[Any] = input_string if string.split(""".""" )[0] == "timestep_embed": return string.replace("""timestep_embed""" , """time_proj""" ) UpperCamelCase : int = 0 if string.startswith("""net.3.""" ): depth += 1 UpperCamelCase : List[str] = string[6:] elif string.startswith("""net.""" ): UpperCamelCase : List[Any] = string[4:] while string.startswith("""main.7.""" ): depth += 1 UpperCamelCase : List[Any] = string[7:] if string.startswith("""main.""" ): UpperCamelCase : Union[str, Any] = string[5:] # mid block if string[:2].isdigit(): UpperCamelCase : List[Any] = string[:2] UpperCamelCase : Optional[Any] = string[2:] else: UpperCamelCase : str = string[0] UpperCamelCase : int = string[1:] if depth == max_depth: UpperCamelCase : List[str] = MID_NUM_TO_LAYER[layer_num] UpperCamelCase : Optional[Any] = """mid_block""" elif depth > 0 and int(SCREAMING_SNAKE_CASE ) < 7: UpperCamelCase : List[str] = DOWN_NUM_TO_LAYER[layer_num] UpperCamelCase : Dict = f"""down_blocks.{depth}""" elif depth > 0 and int(SCREAMING_SNAKE_CASE ) > 7: UpperCamelCase : Any = UP_NUM_TO_LAYER[layer_num] UpperCamelCase : Optional[int] = f"""up_blocks.{max_depth - depth - 1}""" elif depth == 0: UpperCamelCase : str = DEPTH_0_TO_LAYER[layer_num] UpperCamelCase : int = f"""up_blocks.{max_depth - 1}""" if int(SCREAMING_SNAKE_CASE ) > 3 else """down_blocks.0""" if not string_left.startswith(""".""" ): raise ValueError(f"""Naming error with {input_string} and string_left: {string_left}.""" ) UpperCamelCase : Dict = string_left[1:] if "resnets" in new_layer: UpperCamelCase : Optional[Any] = convert_resconv_naming(SCREAMING_SNAKE_CASE ) elif "attentions" in new_layer: UpperCamelCase : Tuple = convert_attn_naming(SCREAMING_SNAKE_CASE ) UpperCamelCase : Union[str, Any] = new_string_left if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): UpperCamelCase : Dict = prefix + """.""" + new_layer + """.""" + string_left else: UpperCamelCase : Tuple = [prefix + """.""" + new_layer + """.""" + s for s in string_left] return new_string def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : List[str] = {} for k, v in state_dict.items(): if k.endswith("""kernel""" ): # up- and downsample layers, don't have trainable weights continue UpperCamelCase : List[Any] = rename(SCREAMING_SNAKE_CASE ) # check if we need to transform from Conv => Linear for attention if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): UpperCamelCase : Any = transform_conv_attns(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: UpperCamelCase : int = v return new_state_dict def UpperCamelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): if len(SCREAMING_SNAKE_CASE ) == 1: if len(v.shape ) == 3: # weight UpperCamelCase : Any = v[:, :, 0] else: # bias UpperCamelCase : List[str] = v else: # qkv matrices UpperCamelCase : Union[str, Any] = v.shape[0] UpperCamelCase : int = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: UpperCamelCase : Tuple = v[i * single_shape : (i + 1) * single_shape, :, 0] else: UpperCamelCase : str = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def UpperCamelCase (SCREAMING_SNAKE_CASE ): UpperCamelCase : List[str] = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) UpperCamelCase : Any = args.model_path.split("""/""" )[-1].split(""".""" )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), f"""Make sure to provide one of the official model names {MODELS_MAP.keys()}""" UpperCamelCase : Union[str, Any] = download(SCREAMING_SNAKE_CASE ) UpperCamelCase : Dict = MODELS_MAP[model_name]["""sample_rate"""] UpperCamelCase : Tuple = MODELS_MAP[model_name]["""sample_size"""] UpperCamelCase : List[str] = Object() UpperCamelCase : Union[str, Any] = sample_size UpperCamelCase : Dict = sample_rate UpperCamelCase : Any = 0 UpperCamelCase : int = UNetaDModel(sample_size=SCREAMING_SNAKE_CASE , sample_rate=SCREAMING_SNAKE_CASE ) UpperCamelCase : Any = diffusers_model.state_dict() UpperCamelCase : Optional[Any] = DiffusionUncond(SCREAMING_SNAKE_CASE ) orig_model.load_state_dict(torch.load(args.model_path , map_location=SCREAMING_SNAKE_CASE )["""state_dict"""] ) UpperCamelCase : Optional[Any] = orig_model.diffusion_ema.eval() UpperCamelCase : Optional[int] = orig_model.state_dict() UpperCamelCase : str = rename_orig_weights(SCREAMING_SNAKE_CASE ) UpperCamelCase : Optional[Any] = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) UpperCamelCase : Optional[Any] = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(SCREAMING_SNAKE_CASE ) == 0, f"""Problem with {renamed_minus_diffusers}""" assert all(k.endswith("""kernel""" ) for k in list(SCREAMING_SNAKE_CASE ) ), f"""Problem with {diffusers_minus_renamed}""" for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), f"""Shape for {key} doesn't match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}""" if key == "time_proj.weight": UpperCamelCase : int = value.squeeze() UpperCamelCase : str = value diffusers_model.load_state_dict(SCREAMING_SNAKE_CASE ) UpperCamelCase : Optional[int] = 100 UpperCamelCase : List[Any] = 33 UpperCamelCase : Tuple = IPNDMScheduler(num_train_timesteps=SCREAMING_SNAKE_CASE ) UpperCamelCase : Optional[int] = torch.manual_seed(SCREAMING_SNAKE_CASE ) UpperCamelCase : List[str] = torch.randn([1, 2, config.sample_size] , generator=SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) UpperCamelCase : str = torch.linspace(1 , 0 , steps + 1 , device=SCREAMING_SNAKE_CASE )[:-1] UpperCamelCase : Tuple = get_crash_schedule(SCREAMING_SNAKE_CASE ) UpperCamelCase : Union[str, Any] = DanceDiffusionPipeline(unet=SCREAMING_SNAKE_CASE , scheduler=SCREAMING_SNAKE_CASE ) UpperCamelCase : Optional[Any] = torch.manual_seed(33 ) UpperCamelCase : Dict = pipe(num_inference_steps=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE ).audios UpperCamelCase : Optional[Any] = sampling.iplms_sample(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , {} ) UpperCamelCase : Any = generated.clamp(-1 , 1 ) UpperCamelCase : List[Any] = (generated - audio).abs().sum() UpperCamelCase : Optional[int] = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print("""Diff sum""" , SCREAMING_SNAKE_CASE ) print("""Diff max""" , SCREAMING_SNAKE_CASE ) assert diff_max < 1E-3, f"""Diff max: {diff_max} is too much :-/""" print(f"""Conversion for {model_name} successful!""" ) if __name__ == "__main__": __magic_name__ : Tuple = argparse.ArgumentParser() parser.add_argument("""--model_path""", default=None, type=str, required=True, help="""Path to the model to convert.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument("""--checkpoint_path""", default=None, type=str, required=True, help="""Path to the output model.""") __magic_name__ : List[str] = parser.parse_args() main(args)
102
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.17.0.dev0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/text-classification/requirements.txt") SCREAMING_SNAKE_CASE : Dict = logging.getLogger(__name__) @dataclass class _lowerCamelCase: lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) lowercase_ : Optional[str] = field( default="""tab_fact""", metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""}, ) lowercase_ : int = field( default=10_24, metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) }, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Whether to pad all samples to `max_seq_length`. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[int] = field( default=_a, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of prediction examples to this """ """value if set.""" ) }, ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the training data."""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """A csv or a json file containing the validation data."""} ) lowercase_ : Optional[str] = field(default=_a, metadata={"""help""": """A csv or a json file containing the test data."""} ) def UpperCamelCase ( self) -> Dict: """simple docstring""" if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError('Need either a GLUE task, a training/validation file or a dataset name.') else: _lowercase : int = self.train_file.split('.')[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." _lowercase : Tuple = self.validation_file.split('.')[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class _lowerCamelCase: lowercase_ : str = field( default=_a, metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowercase_ : Optional[str] = field( default=_a, metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""}, ) lowercase_ : bool = field( default=_a, metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""}, ) lowercase_ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) lowercase_ : bool = field( default=_a, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) def UpperCamelCase_( ) -> Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowercase , _lowercase , _lowercase : Tuple = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowercase , _lowercase , _lowercase : Union[str, Any] = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) _lowercase : Union[str, Any] = training_args.get_process_log_level() logger.setLevel(lowerCamelCase_ ) datasets.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.set_verbosity(lowerCamelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _lowercase : Optional[int] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowercase : Dict = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. _lowercase : Dict = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. _lowercase : Optional[Any] = {'train': data_args.train_file, 'validation': data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: _lowercase : Tuple = data_args.train_file.split('.' )[-1] _lowercase : int = data_args.test_file.split('.' )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." _lowercase : Any = data_args.test_file else: raise ValueError('Need either a GLUE task or a test file for `do_predict`.' ) for key in data_files.keys(): logger.info(F'''load a local file for {key}: {data_files[key]}''' ) if data_args.train_file.endswith('.csv' ): # Loading a dataset from local csv files _lowercase : str = load_dataset('csv' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files _lowercase : Optional[int] = load_dataset('json' , data_files=lowerCamelCase_ , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels _lowercase : Optional[Any] = raw_datasets['train'].features['label'].names _lowercase : Any = len(lowerCamelCase_ ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer _lowercase : str = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=lowerCamelCase_ , ) _lowercase : Tuple = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowerCamelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: _lowercase : int = 'max_length' else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _lowercase : str = False # Some models have set the order of the labels to use, so let's make sure we do use it. _lowercase : List[Any] = {'Refused': 0, 'Entailed': 1} _lowercase : Union[str, Any] = {0: 'Refused', 1: 'Entailed'} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) _lowercase : List[str] = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(lowerCamelCase_ ): # Tokenize the texts def _convert_table_text_to_pandas(lowerCamelCase_ ): _lowercase : int = [_table_row.split('#' ) for _table_row in _table_text.strip('\n' ).split('\n' )] _lowercase : Any = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd _lowercase : List[Any] = examples['statement'] _lowercase : Optional[Any] = list(map(_convert_table_text_to_pandas , examples['table_text'] ) ) _lowercase : Union[str, Any] = tokenizer(lowerCamelCase_ , lowerCamelCase_ , padding=lowerCamelCase_ , max_length=lowerCamelCase_ , truncation=lowerCamelCase_ ) _lowercase : Any = examples['label'] return result with training_args.main_process_first(desc='dataset map pre-processing' ): _lowercase : str = raw_datasets.map( lowerCamelCase_ , batched=lowerCamelCase_ , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on dataset' , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) _lowercase : Any = raw_datasets['train'] if data_args.max_train_samples is not None: _lowercase : str = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) _lowercase : str = raw_datasets['validation'] if data_args.max_eval_samples is not None: _lowercase : List[Any] = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError('--do_predict requires a test dataset' ) _lowercase : Optional[int] = raw_datasets['test'] if data_args.max_predict_samples is not None: _lowercase : List[str] = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(lowerCamelCase_ ) ) , 3 ): logger.info(F'''Sample {index} of the training set: {train_dataset[index]}.''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCamelCase_ ): _lowercase : Dict = p.predictions[0] if isinstance(p.predictions , lowerCamelCase_ ) else p.predictions _lowercase : Tuple = np.argmax(lowerCamelCase_ , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _lowercase : Any = default_data_collator elif training_args.fpaa: _lowercase : str = DataCollatorWithPadding(lowerCamelCase_ , pad_to_multiple_of=8 ) else: _lowercase : Optional[Any] = None # Initialize our Trainer _lowercase : List[str] = Trainer( model=lowerCamelCase_ , args=lowerCamelCase_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=lowerCamelCase_ , tokenizer=lowerCamelCase_ , data_collator=lowerCamelCase_ , ) # Training if training_args.do_train: _lowercase : Optional[int] = None if training_args.resume_from_checkpoint is not None: _lowercase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowercase : Optional[Any] = last_checkpoint _lowercase : Optional[Any] = trainer.train(resume_from_checkpoint=lowerCamelCase_ ) _lowercase : List[Any] = train_result.metrics _lowercase : Dict = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCamelCase_ ) ) _lowercase : int = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics('train' , lowerCamelCase_ ) trainer.save_metrics('train' , lowerCamelCase_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _lowercase : Tuple = trainer.evaluate(eval_dataset=lowerCamelCase_ ) _lowercase : Any = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCamelCase_ ) _lowercase : Optional[int] = min(lowerCamelCase_ , len(lowerCamelCase_ ) ) trainer.log_metrics('eval' , lowerCamelCase_ ) trainer.save_metrics('eval' , lowerCamelCase_ ) if training_args.do_predict: logger.info('*** Predict ***' ) # Removing the `label` columns because it contains -1 and Trainer won't like that. _lowercase : Any = predict_dataset.remove_columns('label' ) _lowercase : Optional[Any] = trainer.predict(lowerCamelCase_ , metric_key_prefix='predict' ).predictions _lowercase : Union[str, Any] = np.argmax(lowerCamelCase_ , axis=1 ) _lowercase : Dict = os.path.join(training_args.output_dir , 'predict_results_tabfact.txt' ) if trainer.is_world_process_zero(): with open(lowerCamelCase_ , 'w' ) as writer: logger.info('***** Predict Results *****' ) writer.write('index\tprediction\n' ) for index, item in enumerate(lowerCamelCase_ ): _lowercase : List[str] = label_list[item] writer.write(F'''{index}\t{item}\n''' ) _lowercase : str = {'finetuned_from': model_args.model_name_or_path, 'tasks': 'text-classification'} if training_args.push_to_hub: trainer.push_to_hub(**lowerCamelCase_ ) else: trainer.create_model_card(**lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ ) -> Dict: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
89
0
"""simple docstring""" import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE,unittest.TestCase ): A__ : List[Any] = RoFormerTokenizer A__ : Tuple = RoFormerTokenizerFast A__ : Union[str, Any] = True A__ : Optional[Any] = True def __UpperCAmelCase ( self : Union[str, Any] ): """simple docstring""" super().setUp() def __UpperCAmelCase ( self : Any , **__lowerCamelCase : Optional[Any] ): """simple docstring""" return self.tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__lowerCamelCase ) def __UpperCAmelCase ( self : Dict , **__lowerCamelCase : List[str] ): """simple docstring""" return self.rust_tokenizer_class.from_pretrained('''junnyu/roformer_chinese_base''' , **__lowerCamelCase ) def __UpperCAmelCase ( self : Any ): """simple docstring""" _snake_case = '''永和服装饰品有限公司,今天天气非常好''' _snake_case = '''永和 服装 饰品 有限公司 , 今 天 天 气 非常 好''' return input_text, output_text def __UpperCAmelCase ( self : List[str] ): """simple docstring""" _snake_case = self.get_tokenizer() _snake_case , _snake_case = self.get_chinese_input_output_texts() _snake_case = tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , output_text.split() ) _snake_case = tokens + [tokenizer.unk_token] _snake_case = [2_2_9_4_3, 2_1_3_3_2, 3_4_4_3_1, 4_5_9_0_4, 1_1_7, 3_0_6, 1_2_3_1, 1_2_3_1, 2_6_5_3, 3_3_9_9_4, 1_2_6_6, 1_0_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , __lowerCamelCase ) def __UpperCAmelCase ( self : Any ): """simple docstring""" _snake_case = self.get_rust_tokenizer() _snake_case , _snake_case = self.get_chinese_input_output_texts() _snake_case = tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , output_text.split() ) _snake_case = tokens + [tokenizer.unk_token] _snake_case = [2_2_9_4_3, 2_1_3_3_2, 3_4_4_3_1, 4_5_9_0_4, 1_1_7, 3_0_6, 1_2_3_1, 1_2_3_1, 2_6_5_3, 3_3_9_9_4, 1_2_6_6, 1_0_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , __lowerCamelCase ) def __UpperCAmelCase ( self : int ): """simple docstring""" pass def __UpperCAmelCase ( self : str ): """simple docstring""" pass def __UpperCAmelCase ( self : Any ): """simple docstring""" pass
103
from maths.prime_factors import prime_factors def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): _lowercase : str = F'''Input value of [number={number}] must be an integer''' raise TypeError(lowerCamelCase_ ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(lowerCamelCase_ ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
89
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { """configuration_mobilenet_v2""": [ """MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileNetV2Config""", """MobileNetV2OnnxConfig""", ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["""MobileNetV2FeatureExtractor"""] UpperCamelCase = ["""MobileNetV2ImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ """MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileNetV2ForImageClassification""", """MobileNetV2ForSemanticSegmentation""", """MobileNetV2Model""", """MobileNetV2PreTrainedModel""", """load_tf_weights_in_mobilenet_v2""", ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
104
from __future__ import annotations from typing import Any class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase, lowerCamelCase = 0) -> None: """simple docstring""" _lowercase , _lowercase : str = row, column _lowercase : Any = [[default_value for c in range(lowerCamelCase)] for r in range(lowerCamelCase)] def __str__( self) -> str: """simple docstring""" _lowercase : Tuple = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _lowercase : str = 0 for row_vector in self.array: for obj in row_vector: _lowercase : Optional[int] = max(lowerCamelCase, len(str(lowerCamelCase))) _lowercase : List[str] = F'''%{max_element_length}s''' # Make string and return def single_line(lowerCamelCase) -> str: nonlocal string_format_identifier _lowercase : Union[str, Any] = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector) line += "]" return line s += "\n".join(single_line(lowerCamelCase) for row_vector in self.array) return s def __repr__( self) -> str: """simple docstring""" return str(self) def UpperCamelCase ( self, lowerCamelCase) -> bool: """simple docstring""" if not (isinstance(lowerCamelCase, (list, tuple)) and len(lowerCamelCase) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self, lowerCamelCase) -> Any: """simple docstring""" assert self.validate_indicies(lowerCamelCase) return self.array[loc[0]][loc[1]] def __setitem__( self, lowerCamelCase, lowerCamelCase) -> None: """simple docstring""" assert self.validate_indicies(lowerCamelCase) _lowercase : Optional[Any] = value def __add__( self, lowerCamelCase) -> Matrix: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) assert self.row == another.row and self.column == another.column # Add _lowercase : Any = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : int = self[r, c] + another[r, c] return result def __neg__( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : List[str] = -self[r, c] return result def __sub__( self, lowerCamelCase) -> Matrix: """simple docstring""" return self + (-another) def __mul__( self, lowerCamelCase) -> Matrix: """simple docstring""" if isinstance(lowerCamelCase, (int, float)): # Scalar multiplication _lowercase : Dict = Matrix(self.row, self.column) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] * another return result elif isinstance(lowerCamelCase, lowerCamelCase): # Matrix multiplication assert self.column == another.row _lowercase : str = Matrix(self.row, another.column) for r in range(self.row): for c in range(another.column): for i in range(self.column): result[r, c] += self[r, i] * another[i, c] return result else: _lowercase : Tuple = F'''Unsupported type given for another ({type(lowerCamelCase)})''' raise TypeError(lowerCamelCase) def UpperCamelCase ( self) -> Matrix: """simple docstring""" _lowercase : List[Any] = Matrix(self.column, self.row) for r in range(self.row): for c in range(self.column): _lowercase : Union[str, Any] = self[r, c] return result def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase) -> Any: """simple docstring""" assert isinstance(lowerCamelCase, lowerCamelCase) and isinstance(lowerCamelCase, lowerCamelCase) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _lowercase : Dict = v.transpose() _lowercase : Any = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def UpperCamelCase_( ) -> None: # a^(-1) _lowercase : Optional[int] = Matrix(3 , 3 , 0 ) for i in range(3 ): _lowercase : int = 1 print(F'''a^(-1) is {ainv}''' ) # u, v _lowercase : Dict = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : Dict = 1, 2, -3 _lowercase : List[Any] = Matrix(3 , 1 , 0 ) _lowercase , _lowercase , _lowercase : int = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCamelCase_ , lowerCamelCase_ )}''' ) def UpperCamelCase_( ) -> None: import doctest doctest.testmod() testa()
89
0
from __future__ import annotations import math def __UpperCAmelCase ( lowerCamelCase_ : int ) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCamelCase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True UpperCamelCase__ : List[str] = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def __UpperCAmelCase ( lowerCamelCase_ : int ) -> list[int]: """simple docstring""" if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) SCREAMING_SNAKE_CASE_ : Dict = [] for num in range(len(lowerCamelCase_ ) ): SCREAMING_SNAKE_CASE_ : List[Any] = 0 while 2 * i * i <= odd_composites[num]: SCREAMING_SNAKE_CASE_ : Dict = odd_composites[num] - 2 * i * i if is_prime(lowerCamelCase_ ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(lowerCamelCase_ ) == n: return list_nums return [] def __UpperCAmelCase ( ) -> int: """simple docstring""" return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
105
import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = int(lowerCamelCase_ ) _lowercase , _lowercase , _lowercase : Optional[Any] = t // 3600, (t // 60) % 60, t % 60 return F'''{h}:{m:02d}:{s:02d}''' if h != 0 else F'''{m:02d}:{s:02d}''' def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=300 ) -> Dict: # docstyle-ignore return F''' <div> {prefix} <progress value=\'{value}\' max=\'{total}\' style=\'width:{width}px; height:20px; vertical-align: middle;\'></progress> {label} </div> ''' def UpperCamelCase_( lowerCamelCase_ ) -> Any: _lowercase : int = '<table border="1" class="dataframe">\n' html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += F''' <th>{i}</th>\n''' html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: _lowercase : Any = F'''{elt:.6f}''' if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else str(lowerCamelCase_ ) html_code += F''' <td>{elt}</td>\n''' html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class _lowerCamelCase: lowercase_ : str = 5 lowercase_ : str = 0.2 def __init__( self, lowerCamelCase, lowerCamelCase = None, lowerCamelCase = True, lowerCamelCase = None, lowerCamelCase = 3_00, ) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = total _lowercase : Optional[int] = '' if prefix is None else prefix _lowercase : Tuple = leave _lowercase : str = parent _lowercase : str = width _lowercase : List[Any] = None _lowercase : List[str] = None _lowercase : Tuple = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, lowerCamelCase = None) -> Dict: """simple docstring""" _lowercase : Any = value if comment is not None: _lowercase : Union[str, Any] = comment if self.last_value is None: _lowercase : Dict = time.time() _lowercase : Tuple = value _lowercase : str = None _lowercase : Optional[int] = self.warmup _lowercase : Optional[Any] = 1 self.update_bar(lowerCamelCase) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for, self.total): if self.first_calls > 0: self.first_calls -= 1 _lowercase : List[str] = time.time() _lowercase : Tuple = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: _lowercase : Dict = self.elapsed_time / (value - self.start_value) else: _lowercase : int = None if value >= self.total: _lowercase : Dict = self.total _lowercase : List[str] = None if not self.leave: self.close() elif self.average_time_per_item is not None: _lowercase : Optional[int] = self.average_time_per_item * (self.total - value) self.update_bar(lowerCamelCase) _lowercase : int = value _lowercase : Tuple = current_time if self.average_time_per_item is None: _lowercase : str = 1 else: _lowercase : int = max(int(self.update_every / self.average_time_per_item), 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None) -> Optional[Any]: """simple docstring""" _lowercase : List[Any] = ' ' * (len(str(self.total)) - len(str(lowerCamelCase))) + str(lowerCamelCase) if self.elapsed_time is None: _lowercase : int = F'''[{spaced_value}/{self.total} : < :''' elif self.predicted_remaining is None: _lowercase : Union[str, Any] = F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)}''' else: _lowercase : Union[str, Any] = ( F'''[{spaced_value}/{self.total} {format_time(self.elapsed_time)} <''' F''' {format_time(self.predicted_remaining)}''' ) self.label += F''', {1/self.average_time_per_item:.2f} it/s''' self.label += "]" if self.comment is None or len(self.comment) == 0 else F''', {self.comment}]''' self.display() def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: _lowercase : Optional[Any] = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" if self.parent is None and self.output is not None: self.output.update(disp.HTML('')) class _lowerCamelCase( _a ): def __init__( self, lowerCamelCase, lowerCamelCase=None) -> int: """simple docstring""" super().__init__(lowerCamelCase) _lowercase : Optional[Any] = None if column_names is None else [column_names] _lowercase : Any = None def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = html_progress_bar(self.value, self.total, self.prefix, self.label, self.width) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: _lowercase : Dict = disp.display(disp.HTML(self.html_code), display_id=lowerCamelCase) else: self.output.update(disp.HTML(self.html_code)) def UpperCamelCase ( self, lowerCamelCase) -> Dict: """simple docstring""" if self.inner_table is None: _lowercase : Dict = [list(values.keys()), list(values.values())] else: _lowercase : Tuple = self.inner_table[0] if len(self.inner_table) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(lowerCamelCase) _lowercase : str = columns self.inner_table.append([values[c] for c in columns]) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=None, lowerCamelCase=3_00) -> Union[str, Any]: """simple docstring""" _lowercase : List[str] = NotebookProgressBar(lowerCamelCase, prefix=lowerCamelCase, parent=self, width=lowerCamelCase) return self.child_bar def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Optional[Any] = None self.display() class _lowerCamelCase( _a ): def __init__( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = None _lowercase : Dict = None _lowercase : Dict = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Dict: """simple docstring""" _lowercase : Dict = 'Epoch' if args.evaluation_strategy == IntervalStrategy.EPOCH else 'Step' _lowercase : Dict = 0 _lowercase : Tuple = 0 _lowercase : int = [self.first_column] + ['Training Loss'] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append('Validation Loss') _lowercase : Union[str, Any] = NotebookTrainingTracker(state.max_steps, lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" _lowercase : Any = int(state.epoch) if int(state.epoch) == state.epoch else F'''{state.epoch:.2f}''' self.training_tracker.update( state.global_step + 1, comment=F'''Epoch {epoch}/{state.num_train_epochs}''', force_update=self._force_next_update, ) _lowercase : str = False def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> Any: """simple docstring""" if not has_length(lowerCamelCase): return if self.prediction_bar is None: if self.training_tracker is not None: _lowercase : Optional[int] = self.training_tracker.add_child(len(lowerCamelCase)) else: _lowercase : Optional[int] = NotebookProgressBar(len(lowerCamelCase)) self.prediction_bar.update(1) else: self.prediction_bar.update(self.prediction_bar.value + 1) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Optional[int]: """simple docstring""" if self.prediction_bar is not None: self.prediction_bar.close() _lowercase : Any = None def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[Any]: """simple docstring""" if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: _lowercase : Dict = {'Training Loss': logs['loss']} # First column is necessarily Step sine we're not in epoch eval strategy _lowercase : List[Any] = state.global_step self.training_tracker.write_line(lowerCamelCase) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, **lowerCamelCase) -> List[str]: """simple docstring""" if self.training_tracker is not None: _lowercase : Tuple = {'Training Loss': 'No log', 'Validation Loss': 'No log'} for log in reversed(state.log_history): if "loss" in log: _lowercase : int = log['loss'] break if self.first_column == "Epoch": _lowercase : Union[str, Any] = int(state.epoch) else: _lowercase : Optional[Any] = state.global_step _lowercase : str = 'eval' for k in metrics: if k.endswith('_loss'): _lowercase : str = re.sub(R'\_loss$', '', lowerCamelCase) _lowercase : Tuple = metrics.pop('total_flos', lowerCamelCase) _lowercase : List[str] = metrics.pop('epoch', lowerCamelCase) _lowercase : List[Any] = metrics.pop(F'''{metric_key_prefix}_runtime''', lowerCamelCase) _lowercase : Dict = metrics.pop(F'''{metric_key_prefix}_samples_per_second''', lowerCamelCase) _lowercase : Tuple = metrics.pop(F'''{metric_key_prefix}_steps_per_second''', lowerCamelCase) _lowercase : List[str] = metrics.pop(F'''{metric_key_prefix}_jit_compilation_time''', lowerCamelCase) for k, v in metrics.items(): if k == F'''{metric_key_prefix}_loss''': _lowercase : Union[str, Any] = v else: _lowercase : Optional[Any] = k.split('_') _lowercase : Optional[int] = ' '.join([part.capitalize() for part in splits[1:]]) _lowercase : Tuple = v self.training_tracker.write_line(lowerCamelCase) self.training_tracker.remove_child() _lowercase : str = None # Evaluation takes a long time so we should force the next update. _lowercase : Optional[Any] = True def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase, **lowerCamelCase) -> Union[str, Any]: """simple docstring""" self.training_tracker.update( state.global_step, comment=F'''Epoch {int(state.epoch)}/{state.num_train_epochs}''', force_update=lowerCamelCase) _lowercase : Any = None
89
0
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging __snake_case :str =logging.get_logger(__name__) class lowerCAmelCase__ ( _lowerCamelCase ): A_ : int = ['input_features', 'attention_mask'] def __init__( self : Tuple , __UpperCamelCase : List[Any]=80 , __UpperCamelCase : int=16_000 , __UpperCamelCase : Tuple=0.0 , __UpperCamelCase : Optional[Any]=10 , __UpperCamelCase : int=25 , __UpperCamelCase : str="hamming_window" , __UpperCamelCase : List[str]=3_2_7_6_8.0 , __UpperCamelCase : Optional[int]=0.9_7 , __UpperCamelCase : List[Any]=1.0 , __UpperCamelCase : int=True , __UpperCamelCase : str=True , __UpperCamelCase : Optional[Any]=False , **__UpperCamelCase : Union[str, Any] , ) -> Tuple: super().__init__(feature_size=__UpperCamelCase , sampling_rate=__UpperCamelCase , padding_value=__UpperCamelCase , **__UpperCamelCase ) A = feature_size A = sampling_rate A = padding_value A = hop_length A = win_length A = frame_signal_scale A = preemphasis_coeff A = mel_floor A = normalize_means A = normalize_vars A = win_function A = return_attention_mask A = win_length * sampling_rate // 1_000 A = hop_length * sampling_rate // 1_000 A = optimal_fft_length(self.sample_size ) A = (self.n_fft // 2) + 1 def __UpperCamelCase ( self : Tuple , __UpperCamelCase : np.array ) -> np.ndarray: if self.win_function == "hamming_window": A = window_function(window_length=self.sample_size , name=self.win_function , periodic=__UpperCamelCase ) else: A = window_function(window_length=self.sample_size , name=self.win_function ) A = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) A = spectrogram( one_waveform * self.frame_signal_scale , window=__UpperCamelCase , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=__UpperCamelCase , preemphasis=self.preemphasis_coeff , mel_filters=__UpperCamelCase , mel_floor=self.mel_floor , log_mel='log' , ) return msfc_features.T def __UpperCamelCase ( self : List[str] , __UpperCamelCase : Optional[Any] , __UpperCamelCase : str , __UpperCamelCase : Tuple ) -> Optional[int]: # make sure we normalize float32 arrays if self.normalize_means: A = x[:input_length].mean(axis=0 ) A = np.subtract(__UpperCamelCase , __UpperCamelCase ) if self.normalize_vars: A = x[:input_length].std(axis=0 ) A = np.divide(__UpperCamelCase , __UpperCamelCase ) if input_length < x.shape[0]: A = padding_value # make sure array is in float32 A = x.astype(np.floataa ) return x def __UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : List[np.ndarray] , __UpperCamelCase : Optional[np.ndarray] = None ) -> List[np.ndarray]: A = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(__UpperCamelCase , __UpperCamelCase , self.padding_value ) for x, n in zip(__UpperCamelCase , __UpperCamelCase )] def __call__( self : Union[str, Any] , __UpperCamelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCamelCase : Union[bool, str, PaddingStrategy] = False , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : bool = False , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : Optional[bool] = None , __UpperCamelCase : Optional[Union[str, TensorType]] = None , __UpperCamelCase : Optional[int] = None , **__UpperCamelCase : List[str] , ) -> BatchFeature: if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) A = isinstance(__UpperCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) A = is_batched_numpy or ( isinstance(__UpperCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: A = [np.asarray(__UpperCamelCase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(__UpperCamelCase , np.ndarray ): A = np.asarray(__UpperCamelCase , dtype=np.floataa ) elif isinstance(__UpperCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): A = raw_speech.astype(np.floataa ) # always return batch if not is_batched: A = [raw_speech] # extract fbank features A = [self._extract_mfsc_features(__UpperCamelCase ) for one_waveform in raw_speech] # convert into correct format for padding A = BatchFeature({'input_features': features} ) A = self.pad( __UpperCamelCase , padding=__UpperCamelCase , max_length=__UpperCamelCase , truncation=__UpperCamelCase , pad_to_multiple_of=__UpperCamelCase , return_attention_mask=__UpperCamelCase , **__UpperCamelCase , ) # make sure list is in array format A = padded_inputs.get('input_features' ) if isinstance(input_features[0] , __UpperCamelCase ): A = [np.asarray(__UpperCamelCase , dtype=np.floataa ) for feature in input_features] A = padded_inputs.get('attention_mask' ) if attention_mask is not None: A = [np.asarray(__UpperCamelCase , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: A = ( np.array(__UpperCamelCase , dtype=np.intaa ) if self._get_padding_strategies(__UpperCamelCase , max_length=__UpperCamelCase ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) A = self.normalize( padded_inputs['input_features'] , attention_mask=__UpperCamelCase ) if return_tensors is not None: A = padded_inputs.convert_to_tensors(__UpperCamelCase ) return padded_inputs
106
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : List[str] = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 18, 2] _lowercase : Tuple = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Any = True if 'large' in model_name or 'huge' in model_name else False _lowercase : Dict = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: _lowercase : Any = [3, 3, 3, 3] _lowercase : Any = [5, 5, 5, 5] elif "fl4" in model_name: _lowercase : Dict = [4, 4, 4, 4] _lowercase : Tuple = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: _lowercase : str = [3, 3, 3, 3] if "lrf" in model_name: _lowercase : Optional[int] = [3, 3, 3, 3] else: _lowercase : Dict = [2, 2, 2, 2] if "tiny" in model_name: _lowercase : List[str] = 96 elif "small" in model_name: _lowercase : Dict = 96 elif "base" in model_name: _lowercase : Optional[int] = 128 elif "large" in model_name: _lowercase : List[Any] = 192 elif "xlarge" in model_name: _lowercase : Optional[Any] = 256 elif "huge" in model_name: _lowercase : Dict = 352 # set label information _lowercase : int = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: _lowercase : str = 'imagenet-22k-id2label.json' else: _lowercase : Tuple = 'imagenet-1k-id2label.json' _lowercase : Union[str, Any] = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : Any = {v: k for k, v in idalabel.items()} _lowercase : Optional[Any] = FocalNetConfig( embed_dim=lowerCamelCase_ , depths=lowerCamelCase_ , focal_levels=lowerCamelCase_ , focal_windows=lowerCamelCase_ , use_conv_embed=lowerCamelCase_ , idalabel=lowerCamelCase_ , labelaid=lowerCamelCase_ , use_post_layernorm=lowerCamelCase_ , use_layerscale=lowerCamelCase_ , ) return config def UpperCamelCase_( lowerCamelCase_ ) -> Any: if "patch_embed.proj" in name: _lowercase : Optional[Any] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: _lowercase : str = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: _lowercase : Any = 'encoder.' + name if "encoder.layers" in name: _lowercase : int = name.replace('encoder.layers' , 'encoder.stages' ) if "downsample.proj" in name: _lowercase : Tuple = name.replace('downsample.proj' , 'downsample.projection' ) if "blocks" in name: _lowercase : str = name.replace('blocks' , 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: _lowercase : List[str] = name.replace('modulation.f' , 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: _lowercase : int = name.replace('modulation.h' , 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: _lowercase : Any = name.replace('modulation.proj' , 'modulation.projection_out' ) if name == "norm.weight": _lowercase : Any = 'layernorm.weight' if name == "norm.bias": _lowercase : Tuple = 'layernorm.bias' if "head" in name: _lowercase : Optional[int] = name.replace('head' , 'classifier' ) else: _lowercase : Optional[int] = 'focalnet.' + name return name def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False ) -> str: # fmt: off _lowercase : Dict = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on _lowercase : Dict = model_name_to_url[model_name] print('Checkpoint URL: ' , lowerCamelCase_ ) _lowercase : List[str] = torch.hub.load_state_dict_from_url(lowerCamelCase_ , map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[int] = val _lowercase : Union[str, Any] = get_focalnet_config(lowerCamelCase_ ) _lowercase : Optional[Any] = FocalNetForImageClassification(lowerCamelCase_ ) model.eval() # load state dict model.load_state_dict(lowerCamelCase_ ) # verify conversion _lowercase : Optional[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : Any = BitImageProcessor( do_resize=lowerCamelCase_ , size={'shortest_edge': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=lowerCamelCase_ , crop_size=224 , do_normalize=lowerCamelCase_ , image_mean=lowerCamelCase_ , image_std=lowerCamelCase_ , ) _lowercase : List[str] = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) _lowercase : List[Any] = processor(images=lowerCamelCase_ , return_tensors='pt' ) _lowercase : str = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ), ] ) _lowercase : List[str] = image_transforms(lowerCamelCase_ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , lowerCamelCase_ , atol=1e-4 ) _lowercase : Dict = model(**lowerCamelCase_ ) _lowercase : int = outputs.logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) print('First values of logits:' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": _lowercase : Optional[Any] = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": _lowercase : int = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": _lowercase : str = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": _lowercase : Any = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": _lowercase : List[Any] = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": _lowercase : int = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3] , lowerCamelCase_ , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class lowercase_ ( unittest.TestCase ): """simple docstring""" @property def __UpperCAmelCase ( self : Any ) -> str: torch.manual_seed(0 ) _A = UNetaDModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=3, out_channels=3, down_block_types=('DownBlock2D', 'AttnDownBlock2D'), up_block_types=('AttnUpBlock2D', 'UpBlock2D'), ) return model def __UpperCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: _A = self.dummy_uncond_unet _A = ScoreSdeVeScheduler() _A = ScoreSdeVePipeline(unet=UpperCamelCase__, scheduler=UpperCamelCase__ ) sde_ve.to(UpperCamelCase__ ) sde_ve.set_progress_bar_config(disable=UpperCamelCase__ ) _A = torch.manual_seed(0 ) _A = sde_ve(num_inference_steps=2, output_type='numpy', generator=UpperCamelCase__ ).images _A = torch.manual_seed(0 ) _A = sde_ve(num_inference_steps=2, output_type='numpy', generator=UpperCamelCase__, return_dict=UpperCamelCase__ )[ 0 ] _A = image[0, -3:, -3:, -1] _A = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _A = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class lowercase_ ( unittest.TestCase ): """simple docstring""" def __UpperCAmelCase ( self : List[str] ) -> List[str]: _A = 'google/ncsnpp-church-256' _A = UNetaDModel.from_pretrained(UpperCamelCase__ ) _A = ScoreSdeVeScheduler.from_pretrained(UpperCamelCase__ ) _A = ScoreSdeVePipeline(unet=UpperCamelCase__, scheduler=UpperCamelCase__ ) sde_ve.to(UpperCamelCase__ ) sde_ve.set_progress_bar_config(disable=UpperCamelCase__ ) _A = torch.manual_seed(0 ) _A = sde_ve(num_inference_steps=10, output_type='numpy', generator=UpperCamelCase__ ).images _A = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) _A = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
107
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Any = { "ut/deta": "https://huggingface.co/ut/deta/resolve/main/config.json", } class _lowerCamelCase( _a ): lowercase_ : Any = """deta""" lowercase_ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self, lowerCamelCase=None, lowerCamelCase=9_00, lowerCamelCase=20_48, lowerCamelCase=6, lowerCamelCase=20_48, lowerCamelCase=8, lowerCamelCase=6, lowerCamelCase=10_24, lowerCamelCase=8, lowerCamelCase=0.0, lowerCamelCase=True, lowerCamelCase="relu", lowerCamelCase=2_56, lowerCamelCase=0.1, lowerCamelCase=0.0, lowerCamelCase=0.0, lowerCamelCase=0.0_2, lowerCamelCase=1.0, lowerCamelCase=True, lowerCamelCase=False, lowerCamelCase="sine", lowerCamelCase=5, lowerCamelCase=4, lowerCamelCase=4, lowerCamelCase=True, lowerCamelCase=3_00, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=1, lowerCamelCase=1, lowerCamelCase=5, lowerCamelCase=2, lowerCamelCase=0.1, lowerCamelCase=0.2_5, **lowerCamelCase, ) -> Any: """simple docstring""" if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.') _lowercase : List[Any] = CONFIG_MAPPING['resnet'](out_features=['stage2', 'stage3', 'stage4']) else: if isinstance(lowerCamelCase, lowerCamelCase): _lowercase : Dict = backbone_config.pop('model_type') _lowercase : int = CONFIG_MAPPING[backbone_model_type] _lowercase : Union[str, Any] = config_class.from_dict(lowerCamelCase) _lowercase : Union[str, Any] = backbone_config _lowercase : Any = num_queries _lowercase : Union[str, Any] = max_position_embeddings _lowercase : Union[str, Any] = d_model _lowercase : Optional[int] = encoder_ffn_dim _lowercase : Optional[int] = encoder_layers _lowercase : Optional[Any] = encoder_attention_heads _lowercase : Optional[Any] = decoder_ffn_dim _lowercase : Dict = decoder_layers _lowercase : Tuple = decoder_attention_heads _lowercase : Union[str, Any] = dropout _lowercase : Optional[Any] = attention_dropout _lowercase : int = activation_dropout _lowercase : Tuple = activation_function _lowercase : List[Any] = init_std _lowercase : Union[str, Any] = init_xavier_std _lowercase : int = encoder_layerdrop _lowercase : Optional[int] = auxiliary_loss _lowercase : Dict = position_embedding_type # deformable attributes _lowercase : Any = num_feature_levels _lowercase : str = encoder_n_points _lowercase : Any = decoder_n_points _lowercase : List[str] = two_stage _lowercase : Dict = two_stage_num_proposals _lowercase : Any = with_box_refine _lowercase : List[Any] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.') # Hungarian matcher _lowercase : List[Any] = class_cost _lowercase : Optional[int] = bbox_cost _lowercase : str = giou_cost # Loss coefficients _lowercase : Optional[int] = mask_loss_coefficient _lowercase : int = dice_loss_coefficient _lowercase : List[Any] = bbox_loss_coefficient _lowercase : Optional[Any] = giou_loss_coefficient _lowercase : str = eos_coefficient _lowercase : int = focal_alpha super().__init__(is_encoder_decoder=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase ( self) -> int: """simple docstring""" return self.d_model def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : int = copy.deepcopy(self.__dict__) _lowercase : Optional[int] = self.backbone_config.to_dict() _lowercase : Optional[Any] = self.__class__.model_type return output
89
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging __a: Optional[Any] = logging.get_logger(__name__) __a: List[str] = { '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/resolve/main/config.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/config.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/config.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _lowerCamelCase = '''bloom''' _lowerCamelCase = ['''past_key_values'''] _lowerCamelCase = { '''num_hidden_layers''': '''n_layer''', '''num_attention_heads''': '''n_head''', } def __init__( self : List[str] , lowerCamelCase : Any=25_0880 , lowerCamelCase : List[str]=64 , lowerCamelCase : Tuple=2 , lowerCamelCase : Any=8 , lowerCamelCase : List[Any]=1E-5 , lowerCamelCase : Any=0.02 , lowerCamelCase : List[str]=True , lowerCamelCase : Any=1 , lowerCamelCase : Any=2 , lowerCamelCase : str=False , lowerCamelCase : List[Any]=0.0 , lowerCamelCase : List[str]=0.0 , lowerCamelCase : Tuple=1 , lowerCamelCase : Dict=False , **lowerCamelCase : Dict , ) -> Any: """simple docstring""" _UpperCAmelCase = vocab_size # Backward compatibility with n_embed kwarg _UpperCAmelCase = kwargs.pop("""n_embed""" , lowerCamelCase ) _UpperCAmelCase = hidden_size if n_embed is None else n_embed _UpperCAmelCase = n_layer _UpperCAmelCase = n_head _UpperCAmelCase = layer_norm_epsilon _UpperCAmelCase = initializer_range _UpperCAmelCase = use_cache _UpperCAmelCase = pretraining_tp _UpperCAmelCase = apply_residual_connection_post_layernorm _UpperCAmelCase = hidden_dropout _UpperCAmelCase = attention_dropout _UpperCAmelCase = bos_token_id _UpperCAmelCase = eos_token_id _UpperCAmelCase = slow_but_exact super().__init__(bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , **lowerCamelCase ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _lowerCamelCase = version.parse('''1.12''' ) def __init__( self : str , lowerCamelCase : PretrainedConfig , lowerCamelCase : str = "default" , lowerCamelCase : List[PatchingSpec] = None , lowerCamelCase : bool = False , ) -> Any: """simple docstring""" super().__init__(lowerCamelCase , task=lowerCamelCase , patching_specs=lowerCamelCase , use_past=lowerCamelCase ) if not getattr(self._config , """pad_token_id""" , lowerCamelCase ): # TODO: how to do that better? _UpperCAmelCase = 0 @property def lowerCamelCase ( self : str ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" _UpperCAmelCase = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(lowerCamelCase , direction="""inputs""" , inverted_values_shape=lowerCamelCase ) _UpperCAmelCase = {0: """batch""", 1: """past_sequence + sequence"""} else: _UpperCAmelCase = {0: """batch""", 1: """sequence"""} return common_inputs @property def lowerCamelCase ( self : str ) -> int: """simple docstring""" return self._config.n_layer @property def lowerCamelCase ( self : Tuple ) -> int: """simple docstring""" return self._config.n_head @property def lowerCamelCase ( self : Any ) -> float: """simple docstring""" return 1E-3 def lowerCamelCase ( self : int , lowerCamelCase : "PreTrainedTokenizer" , lowerCamelCase : int = -1 , lowerCamelCase : int = -1 , lowerCamelCase : bool = False , lowerCamelCase : Optional["TensorType"] = None , ) -> Mapping[str, Any]: """simple docstring""" _UpperCAmelCase = super(lowerCamelCase , self ).generate_dummy_inputs( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) # We need to order the input in the way they appears in the forward() _UpperCAmelCase = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch _UpperCAmelCase , _UpperCAmelCase = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values _UpperCAmelCase = seqlen + 2 _UpperCAmelCase = self._config.hidden_size // self.num_attention_heads _UpperCAmelCase = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) _UpperCAmelCase = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) _UpperCAmelCase = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(self.num_layers ) ] _UpperCAmelCase = common_inputs["""attention_mask"""] if self.use_past: _UpperCAmelCase = ordered_inputs["""attention_mask"""].dtype _UpperCAmelCase = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) return ordered_inputs @property def lowerCamelCase ( self : str ) -> int: """simple docstring""" return 13
108
from __future__ import annotations import numpy as np def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: return np.maximum(0 , lowerCamelCase_ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
89
0
'''simple docstring''' import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation a = logging.get_logger(__name__) a = {"vocab_file": "vocab.txt", "emoji_file": "emoji.json"} a = { "vocab_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt", }, "emoji_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json", }, } a = { "abeja/gpt-neox-japanese-2.7b": 2048, } def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Any: '''simple docstring''' with open(__UpperCAmelCase , """r""" , encoding="""utf-8""" ) as f: __SCREAMING_SNAKE_CASE = json.loads(f.read() ) __SCREAMING_SNAKE_CASE = collections.OrderedDict() __SCREAMING_SNAKE_CASE = collections.OrderedDict() __SCREAMING_SNAKE_CASE = collections.OrderedDict() with open(__UpperCAmelCase , """r""" , encoding="""utf-8""" ) as f: __SCREAMING_SNAKE_CASE = f.readlines() __SCREAMING_SNAKE_CASE = [[t.rstrip("""\n""" )] if (t == """,""" or """,""" not in t) else t.rstrip("""\n""" ).split(""",""" ) for t in token] for idx, b in enumerate(__UpperCAmelCase ): __SCREAMING_SNAKE_CASE = b __SCREAMING_SNAKE_CASE = idx for wd in b: __SCREAMING_SNAKE_CASE = idx return vocab, raw_vocab, ids_to_tokens, emoji class __a ( _snake_case ): __UpperCamelCase : Tuple = VOCAB_FILES_NAMES __UpperCamelCase : Tuple = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self : Dict ,lowerCamelCase : Union[str, Any] ,lowerCamelCase : int ,lowerCamelCase : str="<|endoftext|>" ,lowerCamelCase : List[Any]="<|endoftext|>" ,lowerCamelCase : str="<|startoftext|>" ,lowerCamelCase : Dict="<|endoftext|>" ,lowerCamelCase : int=False ,**lowerCamelCase : str ,): '''simple docstring''' super().__init__( unk_token=lowerCamelCase ,pad_token=lowerCamelCase ,bos_token=lowerCamelCase ,eos_token=lowerCamelCase ,do_clean_text=lowerCamelCase ,**lowerCamelCase ,) if not os.path.isfile(lowerCamelCase ): raise ValueError( f"""Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained""" """ model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) if not os.path.isfile(lowerCamelCase ): raise ValueError( f"""Can't find a emoji file at path '{emoji_file}'. To load the emoji information from a Google""" """ pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) __SCREAMING_SNAKE_CASE = do_clean_text __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = load_vocab_and_emoji(lowerCamelCase ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = SubWordJapaneseTokenizer( vocab=self.vocab ,ids_to_tokens=self.ids_to_tokens ,emoji=self.emoji ) @property def UpperCAmelCase__ ( self : Optional[Any] ): '''simple docstring''' return len(self.raw_vocab ) def UpperCAmelCase__ ( self : List[str] ): '''simple docstring''' return dict(self.raw_vocab ,**self.added_tokens_encoder ) def UpperCAmelCase__ ( self : Dict ,lowerCamelCase : str ): '''simple docstring''' return self.subword_tokenizer.tokenize(lowerCamelCase ,clean=self.do_clean_text ) def UpperCAmelCase__ ( self : Union[str, Any] ,lowerCamelCase : Union[str, Any] ): '''simple docstring''' return self.vocab.get(lowerCamelCase ,self.vocab.get(self.unk_token ) ) def UpperCAmelCase__ ( self : List[Any] ,lowerCamelCase : Optional[Any] ): '''simple docstring''' return self.subword_tokenizer.convert_id_to_token(lowerCamelCase ) def UpperCAmelCase__ ( self : List[str] ,lowerCamelCase : str ): '''simple docstring''' __SCREAMING_SNAKE_CASE = """""".join(lowerCamelCase ).strip() return out_string def UpperCAmelCase__ ( self : List[str] ,lowerCamelCase : "Conversation" ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase ,add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __SCREAMING_SNAKE_CASE = input_ids[-self.model_max_length :] return input_ids def UpperCAmelCase__ ( self : Dict ,lowerCamelCase : str ,lowerCamelCase : Optional[str] = None ): '''simple docstring''' __SCREAMING_SNAKE_CASE = 0 if os.path.isdir(lowerCamelCase ): __SCREAMING_SNAKE_CASE = os.path.join( lowerCamelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __SCREAMING_SNAKE_CASE = os.path.join( lowerCamelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""emoji_file"""] ) else: __SCREAMING_SNAKE_CASE = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""vocab_file"""] ) __SCREAMING_SNAKE_CASE = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""emoji_file"""] ) with open(lowerCamelCase ,"""w""" ,encoding="""utf-8""" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" """ Please check that the vocabulary is not corrupted!""" ) __SCREAMING_SNAKE_CASE = token_index writer.write(""",""".join(lowerCamelCase ) + """\n""" ) index += 1 with open(lowerCamelCase ,"""w""" ,encoding="""utf-8""" ) as writer: json.dump(self.emoji ,lowerCamelCase ) return vocab_file, emoji_file class __a ( _snake_case ): def __init__( self : Tuple ,lowerCamelCase : List[str] ,lowerCamelCase : Union[str, Any] ,lowerCamelCase : Optional[int] ): '''simple docstring''' __SCREAMING_SNAKE_CASE = vocab # same as swe __SCREAMING_SNAKE_CASE = ids_to_tokens # same as bpe __SCREAMING_SNAKE_CASE = emoji __SCREAMING_SNAKE_CASE = np.max([len(lowerCamelCase ) for w in self.vocab.keys()] ) __SCREAMING_SNAKE_CASE = re.compile(r"""(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)""" ) __SCREAMING_SNAKE_CASE = re.compile(r"""[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*""" ) __SCREAMING_SNAKE_CASE = re.compile(r"""[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}""" ) __SCREAMING_SNAKE_CASE = re.compile( r"""([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __SCREAMING_SNAKE_CASE = re.compile( r"""(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __SCREAMING_SNAKE_CASE = re.compile( r"""((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*""" ) __SCREAMING_SNAKE_CASE = """─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿""" __SCREAMING_SNAKE_CASE = """▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟""" __SCREAMING_SNAKE_CASE = str.maketrans({k: """<BLOCK>""" for k in keisen + blocks} ) def __len__( self : List[str] ): '''simple docstring''' return len(self.ids_to_tokens ) def UpperCAmelCase__ ( self : Optional[int] ,lowerCamelCase : Dict ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.content_repattera.sub("""<URL>""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = self.content_repattera.sub("""<EMAIL>""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = self.content_repattera.sub("""<TEL>""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = self.content_repattera.sub("""<DATE>""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = self.content_repattera.sub("""<DATE>""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = self.content_repattera.sub("""<PRICE>""" ,lowerCamelCase ) __SCREAMING_SNAKE_CASE = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __SCREAMING_SNAKE_CASE = content.replace("""<BLOCK><BLOCK>""" ,"""<BLOCK>""" ) return content def UpperCAmelCase__ ( self : Dict ,lowerCamelCase : Optional[Any] ,lowerCamelCase : List[str]=False ): '''simple docstring''' __SCREAMING_SNAKE_CASE = text.replace(""" """ ,"""<SP>""" ) __SCREAMING_SNAKE_CASE = text.replace(""" """ ,"""<SP>""" ) __SCREAMING_SNAKE_CASE = text.replace("""\r\n""" ,"""<BR>""" ) __SCREAMING_SNAKE_CASE = text.replace("""\n""" ,"""<BR>""" ) __SCREAMING_SNAKE_CASE = text.replace("""\r""" ,"""<BR>""" ) __SCREAMING_SNAKE_CASE = text.replace("""\t""" ,"""<TAB>""" ) __SCREAMING_SNAKE_CASE = text.replace("""—""" ,"""ー""" ) __SCREAMING_SNAKE_CASE = text.replace("""−""" ,"""ー""" ) for k, v in self.emoji["emoji"].items(): if k in text: __SCREAMING_SNAKE_CASE = text.replace(lowerCamelCase ,lowerCamelCase ) if clean: __SCREAMING_SNAKE_CASE = self.clean_text(lowerCamelCase ) def check_simbol(lowerCamelCase : List[str] ): __SCREAMING_SNAKE_CASE = x.encode() if len(lowerCamelCase ) == 1 and len(lowerCamelCase ) == 2: __SCREAMING_SNAKE_CASE = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0xC_2A1 and c <= 0xC_2BF) or (c >= 0xC_780 and c <= 0xC_783) or (c >= 0xC_AB9 and c <= 0xC_BBF) or (c >= 0xC_C80 and c <= 0xC_DA2) ): return True return False def checkuae(lowerCamelCase : Union[str, Any] ): __SCREAMING_SNAKE_CASE = x.encode() if len(lowerCamelCase ) == 1 and len(lowerCamelCase ) == 3: __SCREAMING_SNAKE_CASE = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0xE28_080 and c <= 0xE2B_07F: return True return False __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = [] while pos < len(lowerCamelCase ): __SCREAMING_SNAKE_CASE = min(len(lowerCamelCase ) ,pos + self.maxlen + 1 ) if text[pos] == """<""" else pos + 3 __SCREAMING_SNAKE_CASE = [] # (token_id, token, pos) for e in range(lowerCamelCase ,lowerCamelCase ,-1 ): __SCREAMING_SNAKE_CASE = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(lowerCamelCase ) > 2: __SCREAMING_SNAKE_CASE = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(lowerCamelCase ) > 0: # the smallest token_id is adopted __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = sorted(lowerCamelCase ,key=lambda lowerCamelCase : x[0] )[0] result.append(lowerCamelCase ) __SCREAMING_SNAKE_CASE = e else: __SCREAMING_SNAKE_CASE = pos + 1 __SCREAMING_SNAKE_CASE = text[pos:end] if check_simbol(lowerCamelCase ): result.append("""<KIGOU>""" ) elif checkuae(lowerCamelCase ): result.append("""<U2000U2BFF>""" ) else: for i in wd.encode("""utf-8""" ): result.append("""<|byte%d|>""" % i ) __SCREAMING_SNAKE_CASE = end return result def UpperCAmelCase__ ( self : Optional[Any] ,lowerCamelCase : Dict ,lowerCamelCase : Optional[Any]="\n" ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(lowerCamelCase ) > 0: words.append(bytearray(lowerCamelCase ).decode("""utf-8""" ,errors="""replace""" ) ) __SCREAMING_SNAKE_CASE = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["""emoji_inv"""][word] ) elif word == "<SP>": words.append(""" """ ) elif word == "<BR>": words.append(lowerCamelCase ) elif word == "<TAB>": words.append("""\t""" ) elif word == "<BLOCK>": words.append("""▀""" ) elif word == "<KIGOU>": words.append("""ǀ""" ) elif word == "<U2000U2BFF>": words.append("""‖""" ) else: words.append(lowerCamelCase ) if len(lowerCamelCase ) > 0: words.append(bytearray(lowerCamelCase ).decode("""utf-8""" ,errors="""replace""" ) ) __SCREAMING_SNAKE_CASE = """""".join(lowerCamelCase ) return text
109
import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: # Initialise PyTorch model _lowercase : Optional[int] = TaConfig.from_json_file(lowerCamelCase_ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase : Union[str, Any] = TaForConditionalGeneration(lowerCamelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_ta(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
89
0
"""simple docstring""" def lowerCamelCase ( _snake_case ): UpperCAmelCase__ : List[str] = [0] * len(_snake_case ) for i in range(1 ,len(_snake_case ) ): # use last results for better performance - dynamic programming UpperCAmelCase__ : Optional[int] = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: UpperCAmelCase__ : Union[str, Any] = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 UpperCAmelCase__ : Optional[int] = j return prefix_result def lowerCamelCase ( _snake_case ): return max(prefix_function(_snake_case ) ) if __name__ == "__main__": import doctest doctest.testmod()
110
def UpperCamelCase_( lowerCamelCase_ ) -> int: if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: _lowercase : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def UpperCamelCase_( lowerCamelCase_ ) -> int: _lowercase : Tuple = 0 _lowercase : List[str] = 2 while digits < n: index += 1 _lowercase : Optional[int] = len(str(fibonacci(lowerCamelCase_ ) ) ) return index def UpperCamelCase_( lowerCamelCase_ = 1000 ) -> int: return fibonacci_digits_index(lowerCamelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
89
0
def A__ ( lowercase: Optional[int] ) -> int: return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def A__ ( lowercase: List[str] ) -> bool: A : int =0 A : Optional[int] =number while duplicate > 0: A : Optional[int] =divmod(lowerCamelCase_, 10 ) fact_sum += factorial(lowerCamelCase_ ) return fact_sum == number if __name__ == "__main__": print('''Program to check whether a number is a Krisnamurthy Number or not.''') _lowercase : Any =int(input('''Enter number: ''').strip()) print( f'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
305
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Optional[Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConditionalDetrConfig", "ConditionalDetrOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Union[str, Any] = ["ConditionalDetrFeatureExtractor"] SCREAMING_SNAKE_CASE : Optional[Any] = ["ConditionalDetrImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Any = [ "CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST", "ConditionalDetrForObjectDetection", "ConditionalDetrForSegmentation", "ConditionalDetrModel", "ConditionalDetrPreTrainedModel", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
0
def lowerCamelCase_(lowerCamelCase_ ) -> int: assert column_title.isupper() UpperCAmelCase = 0 UpperCAmelCase = len(lowerCamelCase_ ) - 1 UpperCAmelCase = 0 while index >= 0: UpperCAmelCase = (ord(column_title[index] ) - 64) * pow(26 , lowerCamelCase_ ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
323
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function SCREAMING_SNAKE_CASE : Union[str, Any] = 1.0_5457_1817E-34 # unit of ℏ : J * s SCREAMING_SNAKE_CASE : int = 3E8 # unit of c : m * s^-1 def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> dict[str, float]: if (force, area, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if force < 0: raise ValueError('Magnitude of force can not be negative' ) if distance < 0: raise ValueError('Distance can not be negative' ) if area < 0: raise ValueError('Area can not be negative' ) if force == 0: _lowercase : int = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: _lowercase : List[Any] = (240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: _lowercase : List[Any] = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError('One and only one argument must be 0' ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
89
0
import torch from diffusers import DiffusionPipeline class SCREAMING_SNAKE_CASE ( _a ): '''simple docstring''' def __init__( self : Tuple , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Any ): super().__init__() self.register_modules(unet=UpperCAmelCase_ , scheduler=UpperCAmelCase_ ) def __call__( self : Optional[Any] ): SCREAMING_SNAKE_CASE : Union[str, Any] = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) SCREAMING_SNAKE_CASE : int = 1 SCREAMING_SNAKE_CASE : int = self.unet(UpperCAmelCase_ , UpperCAmelCase_ ).sample SCREAMING_SNAKE_CASE : Tuple = self.scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ).prev_sample SCREAMING_SNAKE_CASE : str = scheduler_output - scheduler_output + torch.ones_like(UpperCAmelCase_ ) return result
62
def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) _lowercase : List[str] = 0 _lowercase : Optional[int] = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Any = [int(lowerCamelCase_ ) for i in num_string] _lowercase : List[Any] = 1 for i in range(0 , len(lowerCamelCase_ ) ): total *= numbers[i] _lowercase : Optional[Any] = str(lowerCamelCase_ ) steps += 1 return steps def UpperCamelCase_( lowerCamelCase_ ) -> int: if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) _lowercase : Optional[int] = 0 _lowercase : str = str(lowerCamelCase_ ) while len(lowerCamelCase_ ) != 1: _lowercase : Dict = [int(lowerCamelCase_ ) for i in num_string] _lowercase : Any = 0 for i in range(0 , len(lowerCamelCase_ ) ): total += numbers[i] _lowercase : Dict = str(lowerCamelCase_ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
89
0
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() a =logging.get_logger("""transformers.models.speecht5""") def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> Optional[Any]: hf_model.apply_weight_norm() __lowerCamelCase : Tuple = checkpoint['input_conv.weight_g'] __lowerCamelCase : Any = checkpoint['input_conv.weight_v'] __lowerCamelCase : Tuple = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): __lowerCamelCase : Any = checkpoint[F"upsamples.{i}.1.weight_g"] __lowerCamelCase : Dict = checkpoint[F"upsamples.{i}.1.weight_v"] __lowerCamelCase : Optional[Any] = checkpoint[F"upsamples.{i}.1.bias"] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): __lowerCamelCase : Tuple = checkpoint[F"blocks.{i}.convs1.{j}.1.weight_g"] __lowerCamelCase : str = checkpoint[F"blocks.{i}.convs1.{j}.1.weight_v"] __lowerCamelCase : Dict = checkpoint[F"blocks.{i}.convs1.{j}.1.bias"] __lowerCamelCase : Tuple = checkpoint[F"blocks.{i}.convs2.{j}.1.weight_g"] __lowerCamelCase : str = checkpoint[F"blocks.{i}.convs2.{j}.1.weight_v"] __lowerCamelCase : Tuple = checkpoint[F"blocks.{i}.convs2.{j}.1.bias"] __lowerCamelCase : Union[str, Any] = checkpoint['output_conv.1.weight_g'] __lowerCamelCase : Dict = checkpoint['output_conv.1.weight_v'] __lowerCamelCase : Tuple = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=None , ) -> Tuple: if config_path is not None: __lowerCamelCase : Union[str, Any] = SpeechTaHifiGanConfig.from_pretrained(lowerCamelCase_ ) else: __lowerCamelCase : str = SpeechTaHifiGanConfig() __lowerCamelCase : Any = SpeechTaHifiGan(lowerCamelCase_ ) __lowerCamelCase : Optional[int] = torch.load(lowerCamelCase_ ) load_weights(orig_checkpoint['model']['generator'] , lowerCamelCase_ , lowerCamelCase_ ) __lowerCamelCase : str = np.load(lowerCamelCase_ ) __lowerCamelCase : List[str] = stats[0].reshape(-1 ) __lowerCamelCase : List[str] = stats[1].reshape(-1 ) __lowerCamelCase : List[Any] = torch.from_numpy(lowerCamelCase_ ).float() __lowerCamelCase : int = torch.from_numpy(lowerCamelCase_ ).float() model.save_pretrained(lowerCamelCase_ ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(lowerCamelCase_ ) if __name__ == "__main__": a =argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--stats_path""", required=True, default=None, type=str, help="""Path to stats.npy file""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) a =parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
652
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) def UpperCamelCase_( lowerCamelCase_ ) -> Optional[int]: # initialize config if "resnet-50" in model_name: _lowercase : Union[str, Any] = ResNetConfig.from_pretrained('microsoft/resnet-50' ) elif "resnet-101" in model_name: _lowercase : Optional[Any] = ResNetConfig.from_pretrained('microsoft/resnet-101' ) else: raise ValueError('Model name should include either resnet50 or resnet101' ) _lowercase : Tuple = DetrConfig(use_timm_backbone=lowerCamelCase_ , backbone_config=lowerCamelCase_ ) # set label attributes _lowercase : Any = 'panoptic' in model_name if is_panoptic: _lowercase : List[Any] = 250 else: _lowercase : str = 91 _lowercase : List[Any] = 'huggingface/label-files' _lowercase : Any = 'coco-detection-id2label.json' _lowercase : Tuple = json.load(open(hf_hub_download(lowerCamelCase_ , lowerCamelCase_ , repo_type='dataset' ) , 'r' ) ) _lowercase : int = {int(lowerCamelCase_ ): v for k, v in idalabel.items()} _lowercase : int = idalabel _lowercase : Any = {v: k for k, v in idalabel.items()} return config, is_panoptic def UpperCamelCase_( lowerCamelCase_ ) -> Any: # here we list all keys to be renamed (original name on the left, our name on the right) _lowercase : List[str] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.conv1.weight', 'backbone.conv_encoder.model.embedder.embedder.convolution.weight') ) rename_keys.append(('backbone.0.body.bn1.weight', 'backbone.conv_encoder.model.embedder.embedder.normalization.weight') ) rename_keys.append(('backbone.0.body.bn1.bias', 'backbone.conv_encoder.model.embedder.embedder.normalization.bias') ) rename_keys.append(('backbone.0.body.bn1.running_mean', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_mean') ) rename_keys.append(('backbone.0.body.bn1.running_var', 'backbone.conv_encoder.model.embedder.embedder.normalization.running_var') ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var''', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean''', ) ) rename_keys.append( ( F'''backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var''', F'''backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var''', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''', ) ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ] ) return rename_keys def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Optional[int]: _lowercase : str = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=False ) -> str: _lowercase : Any = '' if is_panoptic: _lowercase : Optional[Any] = 'detr.' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowercase : int = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Tuple = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[str] = in_proj_weight[:256, :] _lowercase : Tuple = in_proj_bias[:256] _lowercase : List[Any] = in_proj_weight[256:512, :] _lowercase : Any = in_proj_bias[256:512] _lowercase : int = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _lowercase : str = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight''' ) _lowercase : Optional[int] = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowercase : Union[str, Any] = in_proj_weight[:256, :] _lowercase : Dict = in_proj_bias[:256] _lowercase : Tuple = in_proj_weight[256:512, :] _lowercase : Dict = in_proj_bias[256:512] _lowercase : str = in_proj_weight[-256:, :] _lowercase : Optional[int] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _lowercase : Tuple = state_dict.pop( F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight''' ) _lowercase : Dict = state_dict.pop(F'''{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _lowercase : List[str] = in_proj_weight_cross_attn[:256, :] _lowercase : Tuple = in_proj_bias_cross_attn[:256] _lowercase : str = in_proj_weight_cross_attn[256:512, :] _lowercase : Union[str, Any] = in_proj_bias_cross_attn[256:512] _lowercase : List[Any] = in_proj_weight_cross_attn[-256:, :] _lowercase : Dict = in_proj_bias_cross_attn[-256:] def UpperCamelCase_( ) -> List[Any]: _lowercase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : str = Image.open(requests.get(lowerCamelCase_ , stream=lowerCamelCase_ ).raw ) return im @torch.no_grad() def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=False ) -> List[Any]: _lowercase , _lowercase : int = get_detr_config(lowerCamelCase_ ) # load original model from torch hub _lowercase : int = { 'detr-resnet-50': 'detr_resnet50', 'detr-resnet-101': 'detr_resnet101', } logger.info(F'''Converting model {model_name}...''' ) _lowercase : Optional[Any] = torch.hub.load('facebookresearch/detr' , model_name_to_original_name[model_name] , pretrained=lowerCamelCase_ ).eval() _lowercase : str = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCamelCase_ ): if is_panoptic: _lowercase : str = 'detr.' + src rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCamelCase_ , is_panoptic=lowerCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowercase : List[Any] = 'detr.model.' if is_panoptic else 'model.' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('detr' ) and not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ) ): _lowercase : Tuple = state_dict.pop(lowerCamelCase_ ) _lowercase : int = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : Optional[Any] = val elif key.startswith('bbox_attention' ) or key.startswith('mask_head' ): continue else: _lowercase : Optional[Any] = state_dict.pop(lowerCamelCase_ ) _lowercase : Union[str, Any] = val else: if not key.startswith('class_labels_classifier' ) and not key.startswith('bbox_predictor' ): _lowercase : Dict = state_dict.pop(lowerCamelCase_ ) _lowercase : List[str] = val # finally, create HuggingFace model and load state dict _lowercase : Optional[Any] = DetrForSegmentation(lowerCamelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCamelCase_ ) model.load_state_dict(lowerCamelCase_ ) model.eval() # verify our conversion on an image _lowercase : str = 'coco_panoptic' if is_panoptic else 'coco_detection' _lowercase : Optional[int] = DetrImageProcessor(format=lowerCamelCase_ ) _lowercase : str = processor(images=prepare_img() , return_tensors='pt' ) _lowercase : Tuple = encoding['pixel_values'] _lowercase : int = detr(lowerCamelCase_ ) _lowercase : Tuple = model(lowerCamelCase_ ) assert torch.allclose(outputs.logits , original_outputs['pred_logits'] , atol=1e-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs['pred_boxes'] , atol=1e-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs['pred_masks'] , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info('Uploading PyTorch model and image processor to the hub...' ) model.push_to_hub(F'''nielsr/{model_name}''' ) processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_name", default="detr-resnet-50", type=str, choices=["detr-resnet-50", "detr-resnet-101"], help="Name of the DETR model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether to push the model to the hub or not.") SCREAMING_SNAKE_CASE : Tuple = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
89
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : int = logging.get_logger(__name__) _lowerCAmelCase : Tuple = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __magic_name__ ( _a ): """simple docstring""" __UpperCamelCase = """xlnet""" __UpperCamelCase = ["""mems"""] __UpperCamelCase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self :str , snake_case :Union[str, Any]=32_000 , snake_case :int=1_024 , snake_case :Optional[int]=24 , snake_case :List[str]=16 , snake_case :Optional[int]=4_096 , snake_case :Optional[Any]="gelu" , snake_case :str=True , snake_case :List[str]="bi" , snake_case :Union[str, Any]=0.02 , snake_case :Dict=1e-12 , snake_case :str=0.1 , snake_case :Union[str, Any]=512 , snake_case :List[str]=None , snake_case :Optional[Any]=True , snake_case :Any=False , snake_case :Dict=False , snake_case :Dict=-1 , snake_case :Tuple=False , snake_case :Tuple="last" , snake_case :List[Any]=True , snake_case :Optional[int]="tanh" , snake_case :Optional[int]=0.1 , snake_case :List[Any]=5 , snake_case :Union[str, Any]=5 , snake_case :Dict=5 , snake_case :List[str]=1 , snake_case :Optional[int]=2 , **snake_case :int , ): '''simple docstring''' A_ : List[Any] = vocab_size A_ : List[Any] = d_model A_ : List[str] = n_layer A_ : List[str] = n_head if d_model % n_head != 0: raise ValueError(f"\'d_model % n_head\' ({d_model % n_head}) should be equal to 0" ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( f"`d_head` ({kwargs['d_head']}) should be equal to `d_model // n_head` ({d_model // n_head})" ) A_ : Optional[Any] = d_model // n_head A_ : List[Any] = ff_activation A_ : int = d_inner A_ : Any = untie_r A_ : Union[str, Any] = attn_type A_ : Optional[Any] = initializer_range A_ : List[Any] = layer_norm_eps A_ : Dict = dropout A_ : List[str] = mem_len A_ : Optional[int] = reuse_len A_ : str = bi_data A_ : List[Any] = clamp_len A_ : List[Any] = same_length A_ : int = summary_type A_ : str = summary_use_proj A_ : Optional[int] = summary_activation A_ : Tuple = summary_last_dropout A_ : Dict = start_n_top A_ : Union[str, Any] = end_n_top A_ : int = bos_token_id A_ : List[Any] = pad_token_id A_ : Optional[int] = eos_token_id if "use_cache" in kwargs: warnings.warn( "The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`" " instead." , snake_case , ) A_ : Dict = kwargs['use_cache'] A_ : str = use_mems_eval A_ : Tuple = use_mems_train super().__init__(pad_token_id=snake_case , bos_token_id=snake_case , eos_token_id=snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' logger.info(f"The model {self.model_type} is one of the few models that has no sequence length limit." ) return -1 @max_position_embeddings.setter def SCREAMING_SNAKE_CASE ( self :Tuple , snake_case :Any ): '''simple docstring''' raise NotImplementedError( f"The model {self.model_type} is one of the few models that has no sequence length limit." )
454
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput SCREAMING_SNAKE_CASE : str = "scheduler_config.json" class _lowerCamelCase( _a ): lowercase_ : Any = 1 lowercase_ : Dict = 2 lowercase_ : Union[str, Any] = 3 lowercase_ : Tuple = 4 lowercase_ : Optional[Any] = 5 @dataclass class _lowerCamelCase( _a ): lowercase_ : jnp.ndarray class _lowerCamelCase: lowercase_ : Union[str, Any] = SCHEDULER_CONFIG_NAME lowercase_ : str = ["""dtype"""] lowercase_ : Dict = [] lowercase_ : int = True @classmethod def UpperCamelCase ( cls, lowerCamelCase = None, lowerCamelCase = None, lowerCamelCase=False, **lowerCamelCase, ) -> Optional[int]: """simple docstring""" _lowercase , _lowercase : Optional[int] = cls.load_config( pretrained_model_name_or_path=lowerCamelCase, subfolder=lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase, ) _lowercase , _lowercase : Tuple = cls.from_config(lowerCamelCase, return_unused_kwargs=lowerCamelCase, **lowerCamelCase) if hasattr(lowerCamelCase, 'create_state') and getattr(lowerCamelCase, 'has_state', lowerCamelCase): _lowercase : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase = False, **lowerCamelCase) -> Any: """simple docstring""" self.save_config(save_directory=lowerCamelCase, push_to_hub=lowerCamelCase, **lowerCamelCase) @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return self._get_compatibles() @classmethod def UpperCamelCase ( cls) -> Any: """simple docstring""" _lowercase : Any = list(set([cls.__name__] + cls._compatibles)) _lowercase : Dict = importlib.import_module(__name__.split('.')[0]) _lowercase : Any = [ getattr(lowerCamelCase, lowerCamelCase) for c in compatible_classes_str if hasattr(lowerCamelCase, lowerCamelCase) ] return compatible_classes def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> jnp.ndarray: assert len(lowerCamelCase_ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(lowerCamelCase_ ) - x.ndim) ) , lowerCamelCase_ ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=0.9_99 , lowerCamelCase_=jnp.floataa ) -> jnp.ndarray: def alpha_bar(lowerCamelCase_ ): return math.cos((time_step + 0.0_08) / 1.0_08 * math.pi / 2 ) ** 2 _lowercase : List[Any] = [] for i in range(lowerCamelCase_ ): _lowercase : Any = i / num_diffusion_timesteps _lowercase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(lowerCamelCase_ ) / alpha_bar(lowerCamelCase_ ) , lowerCamelCase_ ) ) return jnp.array(lowerCamelCase_ , dtype=lowerCamelCase_ ) @flax.struct.dataclass class _lowerCamelCase: lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray lowercase_ : jnp.ndarray @classmethod def UpperCamelCase ( cls, lowerCamelCase) -> str: """simple docstring""" _lowercase : int = scheduler.config if config.trained_betas is not None: _lowercase : str = jnp.asarray(config.trained_betas, dtype=scheduler.dtype) elif config.beta_schedule == "linear": _lowercase : List[Any] = jnp.linspace(config.beta_start, config.beta_end, config.num_train_timesteps, dtype=scheduler.dtype) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _lowercase : Dict = ( jnp.linspace( config.beta_start**0.5, config.beta_end**0.5, config.num_train_timesteps, dtype=scheduler.dtype) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _lowercase : Optional[int] = betas_for_alpha_bar(config.num_train_timesteps, dtype=scheduler.dtype) else: raise NotImplementedError( F'''beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}''') _lowercase : List[str] = 1.0 - betas _lowercase : Union[str, Any] = jnp.cumprod(lowerCamelCase, axis=0) return cls( alphas=lowerCamelCase, betas=lowerCamelCase, alphas_cumprod=lowerCamelCase, ) def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> int: _lowercase : str = state.alphas_cumprod _lowercase : str = alphas_cumprod[timesteps] ** 0.5 _lowercase : Optional[Any] = sqrt_alpha_prod.flatten() _lowercase : Tuple = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) _lowercase : Union[str, Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowercase : Optional[Any] = sqrt_one_minus_alpha_prod.flatten() _lowercase : int = broadcast_to_shape_from_left(lowerCamelCase_ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: _lowercase , _lowercase : Optional[int] = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: _lowercase , _lowercase : Tuple = get_sqrt_alpha_prod(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
89
0
'''simple docstring''' import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed UpperCamelCase_ = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(4_2) UpperCamelCase_ = "sshleifer/student_marian_en_ro_6_1" UpperCamelCase_ = "sshleifer/tiny-mbart" @require_torch class _a ( _a ): '''simple docstring''' def UpperCamelCase_ ( self, A=False, A=None, A=True, A=True, A=True, A=True, ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = self.run_trainer( eval_steps=1, max_len=12, model_name=A, num_train_epochs=1, distributed=A, extra_args_str=A, predict_with_generate=A, do_train=A, do_eval=A, do_predict=A, ) SCREAMING_SNAKE_CASE : List[str] = TrainerState.load_from_json(os.path.join(A, 'trainer_state.json' ) ).log_history if not do_eval: return SCREAMING_SNAKE_CASE : Dict = [log for log in logs if 'eval_loss' in log.keys()] SCREAMING_SNAKE_CASE : List[str] = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats SCREAMING_SNAKE_CASE : Optional[int] = eval_metrics[-1] assert isinstance(last_step_stats['eval_bleu'], A ) assert not math.isnan(float(last_step_stats['eval_loss'] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick() @require_torch_multi_gpu def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick(distributed=A ) @require_torch_multi_gpu def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick(distributed=A ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick(distributed=A, extra_args_str='--sharded_ddp simple' ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick(distributed=A, extra_args_str='--sharded_ddp simple --fp16' ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick(distributed=A, extra_args_str='--sharded_ddp zero_dp_2', predict_with_generate=A ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick( distributed=A, extra_args_str='--sharded_ddp zero_dp_2 --fp16', predict_with_generate=A ) @require_apex @require_torch_gpu def UpperCamelCase_ ( self ): '''simple docstring''' self.run_seqaseq_quick(distributed=A, extra_args_str='--fp16 --fp16_backend=apex' ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=A, extra_args_str='--fp16 --fp16_backend=apex' ) @parameterized.expand(['base', 'low', 'high', 'mixed'] ) @require_torch_multi_gpu def UpperCamelCase_ ( self, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = { # test with the default log_level - should be info and thus log info once 'base': {'extra_args_str': '', 'n_matches': 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes 'low': {'extra_args_str': '--log_level debug --log_level_replica debug', 'n_matches': 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica 'high': {'extra_args_str': '--log_level error --log_level_replica debug', 'n_matches': 1}, # test with high log_level and log_level_replica - should be quiet on all processes 'mixed': {'extra_args_str': '--log_level error --log_level_replica error', 'n_matches': 0}, } SCREAMING_SNAKE_CASE : Tuple = experiments[experiment_id] SCREAMING_SNAKE_CASE : List[str] = {'distributed': True, 'predict_with_generate': False, 'do_eval': False, 'do_predict': False} SCREAMING_SNAKE_CASE : List[Any] = 'Running training' with CaptureStderr() as cl: self.run_seqaseq_quick(**A, extra_args_str=data['extra_args_str'] ) SCREAMING_SNAKE_CASE : List[str] = len(re.findall(A, cl.err ) ) self.assertEqual(A, data['n_matches'] ) @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = self.run_trainer( eval_steps=2, max_len=128, model_name=A, learning_rate=3E-4, num_train_epochs=10, distributed=A, ) # Check metrics SCREAMING_SNAKE_CASE : Optional[Any] = TrainerState.load_from_json(os.path.join(A, 'trainer_state.json' ) ).log_history SCREAMING_SNAKE_CASE : Any = [log for log in logs if 'eval_loss' in log.keys()] SCREAMING_SNAKE_CASE : Union[str, Any] = eval_metrics[0] SCREAMING_SNAKE_CASE : int = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats['eval_bleu'], A ) # test if do_predict saves generations and metrics SCREAMING_SNAKE_CASE : List[Any] = os.listdir(A ) SCREAMING_SNAKE_CASE : str = {os.path.basename(A ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def UpperCamelCase_ ( self ): '''simple docstring''' from transformers.training_args import OptimizerNames def train_and_return_metrics(A ) -> Tuple[int, float]: SCREAMING_SNAKE_CASE : Optional[Any] = '--skip_memory_metrics 0' SCREAMING_SNAKE_CASE : Any = self.run_trainer( max_len=128, model_name=A, learning_rate=3E-4, num_train_epochs=1, optim=A, distributed=A, extra_args_str=A, do_eval=A, do_predict=A, n_gpus_to_use=1, ) # Check metrics SCREAMING_SNAKE_CASE : int = TrainerState.load_from_json(Path(A, 'trainer_state.json' ) ).log_history SCREAMING_SNAKE_CASE : str = int(logs[0]['train_mem_gpu_peaked_delta'] / 2**20 ) SCREAMING_SNAKE_CASE : Optional[int] = int(logs[0]['train_mem_gpu_alloc_delta'] / 2**20 ) SCREAMING_SNAKE_CASE : int = logs[0]['train_loss'] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss SCREAMING_SNAKE_CASE : Optional[int] = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) SCREAMING_SNAKE_CASE : Dict = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) SCREAMING_SNAKE_CASE : Union[str, Any] = gpu_alloc_mem_orig - gpu_alloc_mem_bnb SCREAMING_SNAKE_CASE : Any = gpu_peak_mem_orig + gpu_alloc_mem_orig SCREAMING_SNAKE_CASE : Dict = gpu_peak_mem_bnb + gpu_alloc_mem_bnb SCREAMING_SNAKE_CASE : Tuple = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings SCREAMING_SNAKE_CASE : List[str] = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( A, A, 'should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got' F" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and" F" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB", ) self.assertGreater( A, A, 'should use ~150MB less total gpu memory with BNB, compared to without it for this model but got' F" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and" F" gpu_total_mem_bnb={gpu_total_mem_bnb}MB", ) self.assertEqual( A, A, F"loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}" ) def UpperCamelCase_ ( self, A, A, A, A = 3E-3, A = "adafactor", A = False, A = None, A = 0, A = True, A = True, A = True, A = True, A = None, ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = self.test_file_dir / '../fixtures/tests_samples/wmt_en_ro' SCREAMING_SNAKE_CASE : int = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE : List[str] = F"\n --model_name_or_path {model_name}\n --train_file {data_dir}/train.json\n --validation_file {data_dir}/val.json\n --test_file {data_dir}/test.json\n --output_dir {output_dir}\n --overwrite_output_dir\n --max_train_samples 8\n --max_source_length {max_len}\n --max_target_length {max_len}\n --do_train\n --num_train_epochs {str(A )}\n --per_device_train_batch_size 4\n --learning_rate {learning_rate}\n --warmup_steps 8\n --logging_steps 0\n --logging_strategy no\n --save_steps {str(A )}\n --group_by_length\n --label_smoothing_factor 0.1\n --target_lang ro_RO\n --source_lang en_XX\n ".split() SCREAMING_SNAKE_CASE : str = F"\n --do_eval\n --per_device_eval_batch_size 4\n --max_eval_samples 8\n --val_max_target_length {max_len}\n --evaluation_strategy steps\n --eval_steps {str(A )}\n ".split() SCREAMING_SNAKE_CASE : Dict = '\n --do_predict\n '.split() SCREAMING_SNAKE_CASE : Optional[int] = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"--optim {optim}".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: SCREAMING_SNAKE_CASE : Any = get_gpu_count() SCREAMING_SNAKE_CASE : List[Any] = get_torch_dist_unique_port() SCREAMING_SNAKE_CASE : str = F"\n -m torch.distributed.run\n --nproc_per_node={n_gpus_to_use}\n --master_port={master_port}\n {self.examples_dir_str}/pytorch/translation/run_translation.py\n ".split() SCREAMING_SNAKE_CASE : List[Any] = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(A, env=self.get_env() ) else: SCREAMING_SNAKE_CASE : int = ['run_translation.py'] + args with patch.object(A, 'argv', A ): main() return output_dir
28
from __future__ import annotations def UpperCamelCase_( lowerCamelCase_ ) -> float: if not nums: raise ValueError('List is empty' ) return sum(lowerCamelCase_ ) / len(lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
89
0
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class UpperCAmelCase : '''simple docstring''' def __init__( self : str ,A : Union[str, Any] ,A : List[Any]=None ,A : str=None ,A : Optional[int]=None ,A : Optional[Any]="resnet50" ,A : List[str]=3 ,A : Dict=32 ,A : Union[str, Any]=3 ,A : List[str]=True ,A : Optional[Any]=True ,): __A = parent __A = out_indices if out_indices is not None else [4] __A = stage_names __A = out_features __A = backbone __A = batch_size __A = image_size __A = num_channels __A = use_pretrained_backbone __A = is_training def UpperCamelCase_ ( self : Optional[Any] ): __A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __A = self.get_config() return config, pixel_values def UpperCamelCase_ ( self : Optional[Any] ): return TimmBackboneConfig( image_size=self.image_size ,num_channels=self.num_channels ,out_features=self.out_features ,out_indices=self.out_indices ,stage_names=self.stage_names ,use_pretrained_backbone=self.use_pretrained_backbone ,backbone=self.backbone ,) def UpperCamelCase_ ( self : Optional[Any] ,A : List[str] ,A : Tuple ): __A = TimmBackbone(config=A ) model.to(A ) model.eval() with torch.no_grad(): __A = model(A ) self.parent.assertEqual( result.feature_map[-1].shape ,(self.batch_size, model.channels[-1], 14, 14) ,) def UpperCamelCase_ ( self : Tuple ): __A = self.prepare_config_and_inputs() __A = config_and_inputs __A = {'pixel_values': pixel_values} return config, inputs_dict @require_torch @require_timm class UpperCAmelCase ( _a , _a , _a , unittest.TestCase ): '''simple docstring''' snake_case_ = (TimmBackbone,) if is_torch_available() else () snake_case_ = {"""feature-extraction""": TimmBackbone} if is_torch_available() else {} snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def UpperCamelCase_ ( self : Dict ): __A = TimmBackboneModelTester(self ) __A = ConfigTester(self ,config_class=A ,has_text_modality=A ) def UpperCamelCase_ ( self : List[str] ): self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self : Dict ): __A = 'resnet18' __A = 'microsoft/resnet-18' __A = AutoBackbone.from_pretrained(A ,use_timm_backbone=A ) __A = AutoBackbone.from_pretrained(A ) self.assertEqual(len(timm_model.out_features ) ,len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) ,len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels ,transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices ,(-1,) ) self.assertEqual(transformers_model.out_indices ,[len(timm_model.stage_names ) - 1] ) __A = AutoBackbone.from_pretrained(A ,use_timm_backbone=A ,out_indices=[1, 2, 3] ) __A = AutoBackbone.from_pretrained(A ,out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices ,transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) ,len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels ,transformers_model.channels ) @unittest.skip("TimmBackbone doesn\'t support feed forward chunking" ) def UpperCamelCase_ ( self : str ): pass @unittest.skip("TimmBackbone doesn\'t have num_hidden_layers attribute" ) def UpperCamelCase_ ( self : Dict ): pass @unittest.skip("TimmBackbone initialization is managed on the timm side" ) def UpperCamelCase_ ( self : Union[str, Any] ): pass @unittest.skip("TimmBackbone models doesn\'t have inputs_embeds" ) def UpperCamelCase_ ( self : str ): pass @unittest.skip("TimmBackbone models doesn\'t have inputs_embeds" ) def UpperCamelCase_ ( self : int ): pass @unittest.skip("TimmBackbone model cannot be created without specifying a backbone checkpoint" ) def UpperCamelCase_ ( self : int ): pass @unittest.skip("Only checkpoints on timm can be loaded into TimmBackbone" ) def UpperCamelCase_ ( self : Optional[Any] ): pass @unittest.skip("model weights aren\'t tied in TimmBackbone." ) def UpperCamelCase_ ( self : str ): pass @unittest.skip("model weights aren\'t tied in TimmBackbone." ) def UpperCamelCase_ ( self : int ): pass @unittest.skip("Only checkpoints on timm can be loaded into TimmBackbone" ) def UpperCamelCase_ ( self : Any ): pass @unittest.skip("Only checkpoints on timm can be loaded into TimmBackbone" ) def UpperCamelCase_ ( self : Any ): pass @unittest.skip("TimmBackbone doesn\'t have hidden size info in its configuration." ) def UpperCamelCase_ ( self : Optional[int] ): pass @unittest.skip("TimmBackbone doesn\'t support output_attentions." ) def UpperCamelCase_ ( self : Optional[int] ): pass @unittest.skip("Safetensors is not supported by timm." ) def UpperCamelCase_ ( self : Union[str, Any] ): pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCamelCase_ ( self : Tuple ): pass def UpperCamelCase_ ( self : Dict ): __A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __A = model_class(A ) __A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __A = [*signature.parameters.keys()] __A = ['pixel_values'] self.assertListEqual(arg_names[:1] ,A ) def UpperCamelCase_ ( self : Optional[int] ): __A = self.model_tester.prepare_config_and_inputs_for_common() __A = True __A = self.has_attentions # no need to test all models as different heads yield the same functionality __A = self.all_model_classes[0] __A = model_class(A ) model.to(A ) __A = self._prepare_for_class(A ,A ) __A = model(**A ) __A = outputs[0][-1] # Encoder-/Decoder-only models __A = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: __A = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=A ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def UpperCamelCase_ ( self : Optional[int] ): __A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __A = model_class(A ) model.to(A ) model.eval() __A = model(**A ) self.assertEqual(len(result.feature_maps ) ,len(config.out_indices ) ) self.assertEqual(len(model.channels ) ,len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None __A = copy.deepcopy(A ) __A = None __A = model_class(A ) model.to(A ) model.eval() __A = model(**A ) self.assertEqual(len(result.feature_maps ) ,1 ) self.assertEqual(len(model.channels ) ,1 ) # Check backbone can be initialized with fresh weights __A = copy.deepcopy(A ) __A = False __A = model_class(A ) model.to(A ) model.eval() __A = model(**A )
55
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def UpperCamelCase_( ) -> List[Any]: _lowercase : int = ArgumentParser('Transformers CLI tool' , usage='transformers-cli <command> [<args>]' ) _lowercase : Optional[Any] = parser.add_subparsers(help='transformers-cli command helpers' ) # Register commands ConvertCommand.register_subcommand(lowerCamelCase_ ) DownloadCommand.register_subcommand(lowerCamelCase_ ) EnvironmentCommand.register_subcommand(lowerCamelCase_ ) RunCommand.register_subcommand(lowerCamelCase_ ) ServeCommand.register_subcommand(lowerCamelCase_ ) UserCommands.register_subcommand(lowerCamelCase_ ) AddNewModelCommand.register_subcommand(lowerCamelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCamelCase_ ) LfsCommands.register_subcommand(lowerCamelCase_ ) PTtoTFCommand.register_subcommand(lowerCamelCase_ ) # Let's go _lowercase : Any = parser.parse_args() if not hasattr(lowerCamelCase_ , 'func' ): parser.print_help() exit(1 ) # Run _lowercase : Optional[int] = args.func(lowerCamelCase_ ) service.run() if __name__ == "__main__": main()
89
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : List[str] =logging.get_logger(__name__) _UpperCamelCase : Optional[int] ={ "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json" ), "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json" ), "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json" ), } class UpperCAmelCase__ ( _a ): __snake_case : Dict = """dpr""" def __init__( self ,A__=30522 ,A__=768 ,A__=12 ,A__=12 ,A__=3072 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=512 ,A__=2 ,A__=0.02 ,A__=1E-12 ,A__=0 ,A__="absolute" ,A__ = 0 ,**A__ ,): super().__init__(pad_token_id=A__ ,**A__ ) _A : Optional[Any] = vocab_size _A : Optional[int] = hidden_size _A : Tuple = num_hidden_layers _A : str = num_attention_heads _A : Tuple = hidden_act _A : Union[str, Any] = intermediate_size _A : Optional[Any] = hidden_dropout_prob _A : int = attention_probs_dropout_prob _A : List[str] = max_position_embeddings _A : Optional[int] = type_vocab_size _A : List[str] = initializer_range _A : Dict = layer_norm_eps _A : str = projection_dim _A : int = position_embedding_type
206
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Optional[int] = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def UpperCamelCase ( self, lowerCamelCase=0) -> str: """simple docstring""" _lowercase : Optional[int] = np.random.RandomState(lowerCamelCase) _lowercase : Union[str, Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : List[Any] = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[Any] = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Dict = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> Dict: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[int] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : List[str] = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : Any = pipe(**lowerCamelCase).images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) _lowercase : Any = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_dummy_inputs() _lowercase : Any = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : Optional[int] = output.images[0, -3:, -3:, -1] _lowercase : int = self.get_dummy_inputs() _lowercase : Union[str, Any] = 3 * [inputs.pop('prompt')] _lowercase : Union[str, Any] = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Tuple = text_inputs['input_ids'] _lowercase : Any = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0] _lowercase : List[Any] = prompt_embeds # forward _lowercase : Union[str, Any] = pipe(**lowerCamelCase) _lowercase : Union[str, Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs() _lowercase : Any = 3 * ['this is a negative prompt'] _lowercase : str = negative_prompt _lowercase : Optional[int] = 3 * [inputs['prompt']] # forward _lowercase : int = pipe(**lowerCamelCase) _lowercase : str = output.images[0, -3:, -3:, -1] _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : str = 3 * [inputs.pop('prompt')] _lowercase : Optional[int] = [] for p in [prompt, negative_prompt]: _lowercase : Tuple = pipe.tokenizer( lowerCamelCase, padding='max_length', max_length=pipe.tokenizer.model_max_length, truncation=lowerCamelCase, return_tensors='np', ) _lowercase : Dict = text_inputs['input_ids'] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa))[0]) _lowercase , _lowercase : str = embeds # forward _lowercase : Dict = pipe(**lowerCamelCase) _lowercase : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : int = ort.SessionOptions() _lowercase : str = False return options def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( 'CompVis/stable-diffusion-v1-4', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'A painting of a squirrel eating a burger' np.random.seed(0) _lowercase : Union[str, Any] = sd_pipe([prompt], guidance_scale=6.0, num_inference_steps=10, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : str = DDIMScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : str = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = 'open neural network exchange' _lowercase : List[Any] = np.random.RandomState(0) _lowercase : Optional[Any] = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Dict = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-v1-5', subfolder='scheduler', revision='onnx') _lowercase : Dict = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', scheduler=lowerCamelCase, safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = 'open neural network exchange' _lowercase : str = np.random.RandomState(0) _lowercase : Dict = sd_pipe([prompt], guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np') _lowercase : Optional[Any] = output.images _lowercase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[Any] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase : List[Any] = 0 def test_callback_fn(lowerCamelCase, lowerCamelCase, lowerCamelCase) -> None: _lowercase : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _lowercase : Any = latents[0, -3:, -3:, -1] _lowercase : Tuple = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _lowercase : List[Any] = latents[0, -3:, -3:, -1] _lowercase : str = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 1E-3 _lowercase : Any = False _lowercase : int = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = 'Andromeda galaxy in a bottle' _lowercase : str = np.random.RandomState(0) pipe( prompt=lowerCamelCase, num_inference_steps=5, guidance_scale=7.5, generator=lowerCamelCase, callback=lowerCamelCase, callback_steps=1, ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5', revision='onnx', safety_checker=lowerCamelCase, feature_extractor=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) assert isinstance(lowerCamelCase, lowerCamelCase) assert pipe.safety_checker is None _lowercase : Optional[int] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase) _lowercase : Any = OnnxStableDiffusionPipeline.from_pretrained(lowerCamelCase) # sanity check that the pipeline still works assert pipe.safety_checker is None _lowercase : List[str] = pipe('example prompt', num_inference_steps=2).images[0] assert image is not None
89
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { "configuration_nllb_moe": [ "NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP", "NllbMoeConfig", ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ "NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST", "NllbMoeForConditionalGeneration", "NllbMoeModel", "NllbMoePreTrainedModel", "NllbMoeTop2Router", "NllbMoeSparseMLP", ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys _snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
580
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : List[Any] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = ["PoolFormerFeatureExtractor"] SCREAMING_SNAKE_CASE : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[Any] = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
89
0
import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase=None ) -> Optional[Any]: '''simple docstring''' lowerCamelCase__: Dict = None if token is not None: lowerCamelCase__: int = {'Accept': 'application/vnd.github+json', 'Authorization': f"""Bearer {token}"""} lowerCamelCase__: Dict = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100""" lowerCamelCase__: Optional[Any] = requests.get(lowerCamelCase_ , headers=lowerCamelCase_ ).json() lowerCamelCase__: str = {} try: job_links.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) lowerCamelCase__: int = math.ceil((result["""total_count"""] - 100) / 100 ) for i in range(lowerCamelCase_ ): lowerCamelCase__: List[str] = requests.get(url + f"""&page={i + 2}""" , headers=lowerCamelCase_ ).json() job_links.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) return job_links except Exception: print(f"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase=None ) -> int: '''simple docstring''' lowerCamelCase__: Optional[int] = None if token is not None: lowerCamelCase__: Union[str, Any] = {'Accept': 'application/vnd.github+json', 'Authorization': f"""Bearer {token}"""} lowerCamelCase__: Dict = f"""https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100""" lowerCamelCase__: List[Any] = requests.get(lowerCamelCase_ , headers=lowerCamelCase_ ).json() lowerCamelCase__: Any = {} try: artifacts.update({artifact["""name"""]: artifact["""archive_download_url"""] for artifact in result["""artifacts"""]} ) lowerCamelCase__: Dict = math.ceil((result["""total_count"""] - 100) / 100 ) for i in range(lowerCamelCase_ ): lowerCamelCase__: Union[str, Any] = requests.get(url + f"""&page={i + 2}""" , headers=lowerCamelCase_ ).json() artifacts.update({artifact["""name"""]: artifact["""archive_download_url"""] for artifact in result["""artifacts"""]} ) return artifacts except Exception: print(f"""Unknown error, could not fetch links:\n{traceback.format_exc()}""" ) return {} def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[int]: '''simple docstring''' lowerCamelCase__: List[Any] = None if token is not None: lowerCamelCase__: str = {'Accept': 'application/vnd.github+json', 'Authorization': f"""Bearer {token}"""} lowerCamelCase__: Dict = requests.get(lowerCamelCase_ , headers=lowerCamelCase_ , allow_redirects=lowerCamelCase_ ) lowerCamelCase__: Union[str, Any] = result.headers['Location'] lowerCamelCase__: Dict = requests.get(lowerCamelCase_ , allow_redirects=lowerCamelCase_ ) lowerCamelCase__: Tuple = os.path.join(lowerCamelCase_ , f"""{artifact_name}.zip""" ) with open(lowerCamelCase_ , """wb""" ) as fp: fp.write(response.content ) def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase=None ) -> List[str]: '''simple docstring''' lowerCamelCase__: List[str] = [] lowerCamelCase__: List[Any] = [] lowerCamelCase__: Optional[Any] = None with zipfile.ZipFile(lowerCamelCase_ ) as z: for filename in z.namelist(): if not os.path.isdir(lowerCamelCase_ ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(lowerCamelCase_ ) as f: for line in f: lowerCamelCase__: str = line.decode("""UTF-8""" ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs lowerCamelCase__: Dict = line[: line.index(""": """ )] lowerCamelCase__: Union[str, Any] = line[line.index(""": """ ) + len(""": """ ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith("""FAILED """ ): # `test` is the test method that failed lowerCamelCase__: Union[str, Any] = line[len("""FAILED """ ) :] failed_tests.append(lowerCamelCase_ ) elif filename == "job_name.txt": lowerCamelCase__: List[Any] = line if len(lowerCamelCase_ ) != len(lowerCamelCase_ ): raise ValueError( f"""`errors` and `failed_tests` should have the same number of elements. Got {len(lowerCamelCase_ )} for `errors` """ f"""and {len(lowerCamelCase_ )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some""" """ problem.""" ) lowerCamelCase__: Dict = None if job_name and job_links: lowerCamelCase__: Tuple = job_links.get(lowerCamelCase_ , lowerCamelCase_ ) # A list with elements of the form (line of error, error, failed test) lowerCamelCase__: List[str] = [x + [y] + [job_link] for x, y in zip(lowerCamelCase_ , lowerCamelCase_ )] return result def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase=None ) -> str: '''simple docstring''' lowerCamelCase__: Optional[int] = [] lowerCamelCase__: Union[str, Any] = [os.path.join(lowerCamelCase_ , lowerCamelCase_ ) for p in os.listdir(lowerCamelCase_ ) if p.endswith(""".zip""" )] for p in paths: errors.extend(get_errors_from_single_artifact(lowerCamelCase_ , job_links=lowerCamelCase_ ) ) return errors def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase=None ) -> List[Any]: '''simple docstring''' lowerCamelCase__: int = Counter() counter.update([x[1] for x in logs] ) lowerCamelCase__: Dict = counter.most_common() lowerCamelCase__: Dict = {} for error, count in counts: if error_filter is None or error not in error_filter: lowerCamelCase__: Dict = {'count': count, 'failed_tests': [(x[2], x[0]) for x in logs if x[1] == error]} lowerCamelCase__: Union[str, Any] = dict(sorted(r.items() , key=lambda _UpperCamelCase : item[1]["count"] , reverse=lowerCamelCase_ ) ) return r def __lowerCAmelCase ( _UpperCamelCase ) -> Dict: '''simple docstring''' lowerCamelCase__: Optional[int] = test.split("""::""" )[0] if test.startswith("""tests/models/""" ): lowerCamelCase__: Tuple = test.split("""/""" )[2] else: lowerCamelCase__: str = None return test def __lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase=None ) -> Optional[Any]: '''simple docstring''' lowerCamelCase__: Dict = [(x[0], x[1], get_model(x[2] )) for x in logs] lowerCamelCase__: Dict = [x for x in logs if x[2] is not None] lowerCamelCase__: Optional[int] = {x[2] for x in logs} lowerCamelCase__: Optional[Any] = {} for test in tests: lowerCamelCase__: Union[str, Any] = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) lowerCamelCase__: Tuple = counter.most_common() lowerCamelCase__: List[str] = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} lowerCamelCase__: Union[str, Any] = sum(error_counts.values() ) if n_errors > 0: lowerCamelCase__: List[str] = {'count': n_errors, 'errors': error_counts} lowerCamelCase__: Optional[Any] = dict(sorted(r.items() , key=lambda _UpperCamelCase : item[1]["count"] , reverse=lowerCamelCase_ ) ) return r def __lowerCAmelCase ( _UpperCamelCase ) -> Any: '''simple docstring''' lowerCamelCase__: int = '| no. | error | status |' lowerCamelCase__: int = '|-:|:-|:-|' lowerCamelCase__: Union[str, Any] = [header, sep] for error in reduced_by_error: lowerCamelCase__: str = reduced_by_error[error]['count'] lowerCamelCase__: Dict = f"""| {count} | {error[:100]} | |""" lines.append(lowerCamelCase_ ) return "\n".join(lowerCamelCase_ ) def __lowerCAmelCase ( _UpperCamelCase ) -> Dict: '''simple docstring''' lowerCamelCase__: str = '| model | no. of errors | major error | count |' lowerCamelCase__: Tuple = '|-:|-:|-:|-:|' lowerCamelCase__: List[Any] = [header, sep] for model in reduced_by_model: lowerCamelCase__: str = reduced_by_model[model]['count'] lowerCamelCase__: Dict = list(reduced_by_model[model]["""errors"""].items() )[0] lowerCamelCase__: Dict = f"""| {model} | {count} | {error[:60]} | {_count} |""" lines.append(lowerCamelCase_ ) return "\n".join(lowerCamelCase_ ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') _lowercase = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) _lowercase = get_job_links(args.workflow_run_id, token=args.token) _lowercase = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: _lowercase = k.find(' / ') _lowercase = k[index + len(' / ') :] _lowercase = v with open(os.path.join(args.output_dir, 'job_links.json'), 'w', encoding='UTF-8') as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) _lowercase = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) _lowercase = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error _lowercase = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors _lowercase = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, 'errors.json'), 'w', encoding='UTF-8') as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) _lowercase = reduce_by_error(errors) _lowercase = reduce_by_model(errors) _lowercase = make_github_table(reduced_by_error) _lowercase = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, 'reduced_by_error.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa) with open(os.path.join(args.output_dir, 'reduced_by_model.txt'), 'w', encoding='UTF-8') as fp: fp.write(sa)
306
import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore SCREAMING_SNAKE_CASE : int = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" SCREAMING_SNAKE_CASE : Dict = [file for file in filepaths if file != file.lower()] if upper_files: print(F"{len(upper_files)} files contain uppercase characters:") print("\n".join(upper_files) + "\n") SCREAMING_SNAKE_CASE : List[Any] = [file for file in filepaths if " " in file] if space_files: print(F"{len(space_files)} files contain space characters:") print("\n".join(space_files) + "\n") SCREAMING_SNAKE_CASE : Any = [file for file in filepaths if "-" in file] if hyphen_files: print(F"{len(hyphen_files)} files contain hyphen characters:") print("\n".join(hyphen_files) + "\n") SCREAMING_SNAKE_CASE : str = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"{len(nodir_files)} files are not in a directory:") print("\n".join(nodir_files) + "\n") SCREAMING_SNAKE_CASE : Tuple = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
89
0