code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class UpperCAmelCase_ : lowerCamelCase : Union[str, Any] = LEDConfig lowerCamelCase : Tuple = {} lowerCamelCase : Any = '''gelu''' def __init__( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict=1_3 , UpperCAmelCase__ : Tuple=7 , UpperCAmelCase__ : List[str]=True , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Optional[int]=9_9 , UpperCAmelCase__ : Tuple=3_2 , UpperCAmelCase__ : Optional[Any]=2 , UpperCAmelCase__ : int=4 , UpperCAmelCase__ : str=3_7 , UpperCAmelCase__ : Optional[int]=0.1 , UpperCAmelCase__ : str=0.1 , UpperCAmelCase__ : Dict=2_0 , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : str=4 , ) -> Optional[Any]: lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = eos_token_id lowerCAmelCase = pad_token_id lowerCAmelCase = bos_token_id lowerCAmelCase = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after lowerCAmelCase = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests lowerCAmelCase = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __UpperCAmelCase ( self : Union[str, Any] ) -> int: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) lowerCAmelCase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) lowerCAmelCase = tf.concat([input_ids, eos_tensor] , axis=1 ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) lowerCAmelCase = prepare_led_inputs_dict(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = tf.concat( [tf.zeros_like(UpperCAmelCase__ )[:, :-1], tf.ones_like(UpperCAmelCase__ )[:, -1:]] , axis=-1 , ) lowerCAmelCase = global_attention_mask return config, inputs_dict def __UpperCAmelCase ( self : Any , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Union[str, Any] ) -> Optional[int]: lowerCAmelCase = TFLEDModel(config=UpperCAmelCase__ ).get_decoder() lowerCAmelCase = inputs_dict['input_ids'] lowerCAmelCase = input_ids[:1, :] lowerCAmelCase = inputs_dict['attention_mask'][:1, :] lowerCAmelCase = 1 # first forward pass lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , use_cache=UpperCAmelCase__ ) lowerCAmelCase , lowerCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCAmelCase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and lowerCAmelCase = tf.concat([input_ids, next_tokens] , axis=-1 ) lowerCAmelCase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )[0] lowerCAmelCase = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , past_key_values=UpperCAmelCase__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice lowerCAmelCase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx] lowerCAmelCase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(UpperCAmelCase__ , UpperCAmelCase__ , rtol=1E-3 ) def a_ ( lowerCamelCase : Any , lowerCamelCase : Tuple , lowerCamelCase : Any , lowerCamelCase : List[Any]=None , lowerCamelCase : Any=None , lowerCamelCase : Union[str, Any]=None , lowerCamelCase : List[str]=None , ): if attention_mask is None: lowerCAmelCase = tf.cast(tf.math.not_equal(lowerCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCAmelCase = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCAmelCase = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCAmelCase = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class UpperCAmelCase_ ( __lowercase , __lowercase , unittest.TestCase ): lowerCamelCase : Optional[Any] = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () lowerCamelCase : List[Any] = (TFLEDForConditionalGeneration,) if is_tf_available() else () lowerCamelCase : str = ( { '''conversational''': TFLEDForConditionalGeneration, '''feature-extraction''': TFLEDModel, '''summarization''': TFLEDForConditionalGeneration, '''text2text-generation''': TFLEDForConditionalGeneration, '''translation''': TFLEDForConditionalGeneration, } if is_tf_available() else {} ) lowerCamelCase : Dict = True lowerCamelCase : List[Any] = False lowerCamelCase : List[Any] = False lowerCamelCase : List[str] = False def __UpperCAmelCase ( self : Any ) -> Dict: lowerCAmelCase = TFLEDModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=UpperCAmelCase__ ) def __UpperCAmelCase ( self : Union[str, Any] ) -> int: self.config_tester.run_common_tests() def __UpperCAmelCase ( self : List[str] ) -> Optional[Any]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*UpperCAmelCase__ ) def __UpperCAmelCase ( self : Dict ) -> Tuple: lowerCAmelCase , lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = tf.zeros_like(inputs_dict['attention_mask'] ) lowerCAmelCase = 2 lowerCAmelCase = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['global_attention_mask'] , ) lowerCAmelCase = True lowerCAmelCase = self.model_tester.seq_length lowerCAmelCase = self.model_tester.encoder_seq_length def check_decoder_attentions_output(UpperCAmelCase__ : Any ): lowerCAmelCase = outputs.decoder_attentions self.assertEqual(len(UpperCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(UpperCAmelCase__ : Optional[Any] ): lowerCAmelCase = [t.numpy() for t in outputs.encoder_attentions] lowerCAmelCase = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(UpperCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(UpperCAmelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: lowerCAmelCase = True lowerCAmelCase = False lowerCAmelCase = False lowerCAmelCase = model_class(UpperCAmelCase__ ) lowerCAmelCase = model(self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) lowerCAmelCase = len(UpperCAmelCase__ ) self.assertEqual(config.output_hidden_states , UpperCAmelCase__ ) check_encoder_attentions_output(UpperCAmelCase__ ) if self.is_encoder_decoder: lowerCAmelCase = model_class(UpperCAmelCase__ ) lowerCAmelCase = model(self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertEqual(config.output_hidden_states , UpperCAmelCase__ ) check_decoder_attentions_output(UpperCAmelCase__ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowerCAmelCase = True lowerCAmelCase = model_class(UpperCAmelCase__ ) lowerCAmelCase = model(self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertEqual(config.output_hidden_states , UpperCAmelCase__ ) check_encoder_attentions_output(UpperCAmelCase__ ) # Check attention is always last and order is fine lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = model_class(UpperCAmelCase__ ) lowerCAmelCase = model(self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(UpperCAmelCase__ ) ) self.assertEqual(model.config.output_hidden_states , UpperCAmelCase__ ) check_encoder_attentions_output(UpperCAmelCase__ ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def __UpperCAmelCase ( self : int ) -> List[Any]: pass def __UpperCAmelCase ( self : int ) -> Any: # TODO: Head-masking not yet implement pass def a_ ( lowerCamelCase : Any ): return tf.constant(lowerCamelCase , dtype=tf.intaa ) __snake_case =1e-4 @slow @require_tf class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: lowerCAmelCase = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here lowerCAmelCase = _long_tensor([5_1_2 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) lowerCAmelCase = _long_tensor([1_2_8 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) lowerCAmelCase = prepare_led_inputs_dict(model.config , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = model(**UpperCAmelCase__ )[0] lowerCAmelCase = (1, 1_0_2_4, 7_6_8) self.assertEqual(output.shape , UpperCAmelCase__ ) # change to expected output here lowerCAmelCase = tf.convert_to_tensor( [[2.3_050, 2.8_279, 0.6_531], [-1.8_457, -0.1_455, -3.5_661], [-1.0_186, 0.4_586, -2.2_043]] , ) tf.debugging.assert_near(output[:, :3, :3] , UpperCAmelCase__ , atol=1E-3 ) def __UpperCAmelCase ( self : List[Any] ) -> Tuple: lowerCAmelCase = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here lowerCAmelCase = _long_tensor([5_1_2 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) lowerCAmelCase = _long_tensor([1_2_8 * [0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9]] ) lowerCAmelCase = prepare_led_inputs_dict(model.config , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = model(**UpperCAmelCase__ )[0] lowerCAmelCase = (1, 1_0_2_4, model.config.vocab_size) self.assertEqual(output.shape , UpperCAmelCase__ ) # change to expected output here lowerCAmelCase = tf.convert_to_tensor( [[33.6_507, 6.4_572, 16.8_089], [5.8_739, -2.4_238, 11.2_902], [-3.2_139, -4.3_149, 4.2_783]] , ) tf.debugging.assert_near(output[:, :3, :3] , UpperCAmelCase__ , atol=1E-3 , rtol=1E-3 )
4
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''google/canine-s''': '''https://huggingface.co/google/canine-s/resolve/main/config.json''', # See all CANINE models at https://huggingface.co/models?filter=canine } class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = '''canine''' def __init__( self : Dict , __UpperCAmelCase : List[str]=768 , __UpperCAmelCase : str=12 , __UpperCAmelCase : Union[str, Any]=12 , __UpperCAmelCase : int=3072 , __UpperCAmelCase : Optional[int]="gelu" , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : str=0.1 , __UpperCAmelCase : List[Any]=16384 , __UpperCAmelCase : Any=16 , __UpperCAmelCase : str=0.02 , __UpperCAmelCase : Dict=1E-12 , __UpperCAmelCase : Optional[Any]=0 , __UpperCAmelCase : int=0xE000 , __UpperCAmelCase : List[Any]=0xE001 , __UpperCAmelCase : Any=4 , __UpperCAmelCase : Dict=4 , __UpperCAmelCase : List[str]=8 , __UpperCAmelCase : int=16384 , __UpperCAmelCase : Union[str, Any]=128 , **__UpperCAmelCase : Dict , ): '''simple docstring''' super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) _A = max_position_embeddings _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = initializer_range _A = type_vocab_size _A = layer_norm_eps # Character config: _A = downsampling_rate _A = upsampling_kernel_size _A = num_hash_functions _A = num_hash_buckets _A = local_transformer_stride
79
0
import os def _A ( ): """simple docstring""" a__ : Optional[int] =os.path.join(os.path.dirname(SCREAMING_SNAKE_CASE ) , "num.txt" ) with open(SCREAMING_SNAKE_CASE ) as file_hand: return str(sum(int(SCREAMING_SNAKE_CASE ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
367
UpperCAmelCase : Dict = [0, 2, 4, 6, 8] UpperCAmelCase : Tuple = [1, 3, 5, 7, 9] def _A ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : int ): """simple docstring""" if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 a__ : str =0 for digit in range(10 ): a__ : int =digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return result a__ : List[str] =0 for digita in range(10 ): a__ : Optional[int] =digita if (remainder + digita) % 2 == 0: a__ : Dict =ODD_DIGITS else: a__ : Any =EVEN_DIGITS for digita in other_parity_digits: a__ : Union[str, Any] =digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) return result def _A ( SCREAMING_SNAKE_CASE : int = 9 ): """simple docstring""" a__ : List[str] =0 for length in range(1 , max_power + 1 ): result += reversible_numbers(SCREAMING_SNAKE_CASE , 0 , [0] * length , SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(F"""{solution() = }""")
148
0
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : int = { "configuration_autoformer": [ "AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "AutoformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = [ "AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "AutoformerForPrediction", "AutoformerModel", "AutoformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
252
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __lowerCamelCase ( lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Tuple ): '''simple docstring''' if gpta_config_file == "": lowerCamelCase = GPTaConfig() else: lowerCamelCase = GPTaConfig.from_json_file(lowerCamelCase__ ) lowerCamelCase = GPTaModel(lowerCamelCase__ ) # Load weights from numpy load_tf_weights_in_gpta(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # Save pytorch-model lowerCamelCase = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME lowerCamelCase = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(f'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , lowerCamelCase__ ) print(f'Save configuration file to {pytorch_config_dump_path}' ) with open(lowerCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": UpperCAmelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--gpt2_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--gpt2_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained OpenAI model. \n" "This specifies the model architecture." ), ) UpperCAmelCase : Tuple = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
252
1
'''simple docstring''' import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) lowerCAmelCase_ : Optional[Any] = { '''sample_size''': 32, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': 1000, '''block_out_channels''': [32, 64], '''attention_head_dim''': 8, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCAmelCase_ : List[str] = { '''sample_size''': 64, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 3, '''num_class_embeds''': 1000, '''block_out_channels''': [192, 192 * 2, 192 * 3, 192 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCAmelCase_ : List[Any] = { '''sample_size''': 256, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': None, '''block_out_channels''': [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''default''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCAmelCase_ : Any = { '''num_train_timesteps''': 40, '''sigma_min''': 0.0_0_2, '''sigma_max''': 8_0.0, } lowerCAmelCase_ : List[str] = { '''num_train_timesteps''': 201, '''sigma_min''': 0.0_0_2, '''sigma_max''': 8_0.0, } lowerCAmelCase_ : Any = { '''num_train_timesteps''': 151, '''sigma_min''': 0.0_0_2, '''sigma_max''': 8_0.0, } def __A ( lowerCAmelCase_ ): if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("""boolean value expected""" ) def __A ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False ): _UpperCAmelCase : int = checkpoint[f"{old_prefix}.in_layers.0.weight"] _UpperCAmelCase : Optional[Any] = checkpoint[f"{old_prefix}.in_layers.0.bias"] _UpperCAmelCase : List[str] = checkpoint[f"{old_prefix}.in_layers.2.weight"] _UpperCAmelCase : str = checkpoint[f"{old_prefix}.in_layers.2.bias"] _UpperCAmelCase : str = checkpoint[f"{old_prefix}.emb_layers.1.weight"] _UpperCAmelCase : str = checkpoint[f"{old_prefix}.emb_layers.1.bias"] _UpperCAmelCase : List[Any] = checkpoint[f"{old_prefix}.out_layers.0.weight"] _UpperCAmelCase : Tuple = checkpoint[f"{old_prefix}.out_layers.0.bias"] _UpperCAmelCase : int = checkpoint[f"{old_prefix}.out_layers.3.weight"] _UpperCAmelCase : Optional[int] = checkpoint[f"{old_prefix}.out_layers.3.bias"] if has_skip: _UpperCAmelCase : Union[str, Any] = checkpoint[f"{old_prefix}.skip_connection.weight"] _UpperCAmelCase : Union[str, Any] = checkpoint[f"{old_prefix}.skip_connection.bias"] return new_checkpoint def __A ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ): _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = checkpoint[f"{old_prefix}.qkv.weight"].chunk(3 , dim=0 ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = checkpoint[f"{old_prefix}.qkv.bias"].chunk(3 , dim=0 ) _UpperCAmelCase : str = checkpoint[f"{old_prefix}.norm.weight"] _UpperCAmelCase : Optional[int] = checkpoint[f"{old_prefix}.norm.bias"] _UpperCAmelCase : Optional[Any] = weight_q.squeeze(-1 ).squeeze(-1 ) _UpperCAmelCase : str = bias_q.squeeze(-1 ).squeeze(-1 ) _UpperCAmelCase : Tuple = weight_k.squeeze(-1 ).squeeze(-1 ) _UpperCAmelCase : Union[str, Any] = bias_k.squeeze(-1 ).squeeze(-1 ) _UpperCAmelCase : List[Any] = weight_v.squeeze(-1 ).squeeze(-1 ) _UpperCAmelCase : List[str] = bias_v.squeeze(-1 ).squeeze(-1 ) _UpperCAmelCase : List[str] = ( checkpoint[f"{old_prefix}.proj_out.weight"].squeeze(-1 ).squeeze(-1 ) ) _UpperCAmelCase : Dict = checkpoint[f"{old_prefix}.proj_out.bias"].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def __A ( lowerCAmelCase_ , lowerCAmelCase_ ): _UpperCAmelCase : List[Any] = torch.load(lowerCAmelCase_ , map_location="""cpu""" ) _UpperCAmelCase : Any = {} _UpperCAmelCase : Union[str, Any] = checkpoint["""time_embed.0.weight"""] _UpperCAmelCase : Any = checkpoint["""time_embed.0.bias"""] _UpperCAmelCase : int = checkpoint["""time_embed.2.weight"""] _UpperCAmelCase : str = checkpoint["""time_embed.2.bias"""] if unet_config["num_class_embeds"] is not None: _UpperCAmelCase : Tuple = checkpoint["""label_emb.weight"""] _UpperCAmelCase : List[Any] = checkpoint["""input_blocks.0.0.weight"""] _UpperCAmelCase : Union[str, Any] = checkpoint["""input_blocks.0.0.bias"""] _UpperCAmelCase : Optional[int] = unet_config["""down_block_types"""] _UpperCAmelCase : List[Any] = unet_config["""layers_per_block"""] _UpperCAmelCase : Optional[int] = unet_config["""attention_head_dim"""] _UpperCAmelCase : List[Any] = unet_config["""block_out_channels"""] _UpperCAmelCase : List[Any] = 1 _UpperCAmelCase : Dict = channels_list[0] for i, layer_type in enumerate(lowerCAmelCase_ ): _UpperCAmelCase : Any = channels_list[i] _UpperCAmelCase : Dict = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(lowerCAmelCase_ ): _UpperCAmelCase : Union[str, Any] = f"down_blocks.{i}.resnets.{j}" _UpperCAmelCase : Any = f"input_blocks.{current_layer}.0" _UpperCAmelCase : Optional[int] = True if j == 0 and downsample_block_has_skip else False _UpperCAmelCase : Optional[int] = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , has_skip=lowerCAmelCase_ ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(lowerCAmelCase_ ): _UpperCAmelCase : Dict = f"down_blocks.{i}.resnets.{j}" _UpperCAmelCase : str = f"input_blocks.{current_layer}.0" _UpperCAmelCase : Tuple = True if j == 0 and downsample_block_has_skip else False _UpperCAmelCase : Dict = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , has_skip=lowerCAmelCase_ ) _UpperCAmelCase : Dict = f"down_blocks.{i}.attentions.{j}" _UpperCAmelCase : Optional[Any] = f"input_blocks.{current_layer}.1" _UpperCAmelCase : str = convert_attention( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) current_layer += 1 if i != len(lowerCAmelCase_ ) - 1: _UpperCAmelCase : Union[str, Any] = f"down_blocks.{i}.downsamplers.0" _UpperCAmelCase : str = f"input_blocks.{current_layer}.0" _UpperCAmelCase : int = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) current_layer += 1 _UpperCAmelCase : str = current_channels # hardcoded the mid-block for now _UpperCAmelCase : Tuple = """mid_block.resnets.0""" _UpperCAmelCase : Optional[Any] = """middle_block.0""" _UpperCAmelCase : Tuple = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : List[Any] = """mid_block.attentions.0""" _UpperCAmelCase : Union[str, Any] = """middle_block.1""" _UpperCAmelCase : Optional[Any] = convert_attention(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : List[str] = """mid_block.resnets.1""" _UpperCAmelCase : Optional[int] = """middle_block.2""" _UpperCAmelCase : Tuple = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : Tuple = unet_config["""up_block_types"""] for i, layer_type in enumerate(lowerCAmelCase_ ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): _UpperCAmelCase : Tuple = f"up_blocks.{i}.resnets.{j}" _UpperCAmelCase : Union[str, Any] = f"output_blocks.{current_layer}.0" _UpperCAmelCase : Optional[Any] = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , has_skip=lowerCAmelCase_ ) current_layer += 1 if i != len(lowerCAmelCase_ ) - 1: _UpperCAmelCase : str = f"up_blocks.{i}.upsamplers.0" _UpperCAmelCase : Dict = f"output_blocks.{current_layer-1}.1" _UpperCAmelCase : Optional[Any] = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): _UpperCAmelCase : Any = f"up_blocks.{i}.resnets.{j}" _UpperCAmelCase : Any = f"output_blocks.{current_layer}.0" _UpperCAmelCase : List[Any] = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , has_skip=lowerCAmelCase_ ) _UpperCAmelCase : str = f"up_blocks.{i}.attentions.{j}" _UpperCAmelCase : Any = f"output_blocks.{current_layer}.1" _UpperCAmelCase : Optional[Any] = convert_attention( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) current_layer += 1 if i != len(lowerCAmelCase_ ) - 1: _UpperCAmelCase : Dict = f"up_blocks.{i}.upsamplers.0" _UpperCAmelCase : Optional[Any] = f"output_blocks.{current_layer-1}.2" _UpperCAmelCase : Dict = convert_resnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : Any = checkpoint["""out.0.weight"""] _UpperCAmelCase : int = checkpoint["""out.0.bias"""] _UpperCAmelCase : List[str] = checkpoint["""out.2.weight"""] _UpperCAmelCase : Tuple = checkpoint["""out.2.bias"""] return new_checkpoint if __name__ == "__main__": lowerCAmelCase_ : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--unet_path''', default=None, type=str, required=True, help='''Path to the unet.pt to convert.''') parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output the converted UNet model.''' ) parser.add_argument('''--class_cond''', default=True, type=str, help='''Whether the model is class-conditional.''') lowerCAmelCase_ : Any = parser.parse_args() lowerCAmelCase_ : Tuple = strabool(args.class_cond) lowerCAmelCase_ : Optional[int] = os.path.basename(args.unet_path) print(F"Checkpoint: {ckpt_name}") # Get U-Net config if "imagenet64" in ckpt_name: lowerCAmelCase_ : Optional[int] = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowerCAmelCase_ : int = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: lowerCAmelCase_ : int = TEST_UNET_CONFIG else: raise ValueError(F"Checkpoint type {ckpt_name} is not currently supported.") if not args.class_cond: lowerCAmelCase_ : List[str] = None lowerCAmelCase_ : List[Any] = con_pt_to_diffuser(args.unet_path, unet_config) lowerCAmelCase_ : List[str] = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: lowerCAmelCase_ : int = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: lowerCAmelCase_ : Dict = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowerCAmelCase_ : Union[str, Any] = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(F"Checkpoint type {ckpt_name} is not currently supported.") lowerCAmelCase_ : Dict = CMStochasticIterativeScheduler(**scheduler_config) lowerCAmelCase_ : Dict = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
170
'''simple docstring''' from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. lowerCAmelCase_ : Optional[Any] = 10 def __A ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): for i in range(lowerCAmelCase_ , lowerCAmelCase_ ): if array[i] == target: return i return -1 def __A ( lowerCAmelCase_ , lowerCAmelCase_ ): _UpperCAmelCase : Optional[int] = 0 _UpperCAmelCase : str = len(lowerCAmelCase_ ) while left <= right: if right - left < precision: return lin_search(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : Tuple = (left + right) // 3 + 1 _UpperCAmelCase : str = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: _UpperCAmelCase : List[Any] = one_third - 1 elif array[two_third] < target: _UpperCAmelCase : Optional[Any] = two_third + 1 else: _UpperCAmelCase : Dict = one_third + 1 _UpperCAmelCase : List[Any] = two_third - 1 else: return -1 def __A ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): if left < right: if right - left < precision: return lin_search(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : List[Any] = (left + right) // 3 + 1 _UpperCAmelCase : int = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(lowerCAmelCase_ , one_third - 1 , lowerCAmelCase_ , lowerCAmelCase_ ) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) else: return rec_ternary_search(one_third + 1 , two_third - 1 , lowerCAmelCase_ , lowerCAmelCase_ ) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase_ : int = input('''Enter numbers separated by comma:\n''').strip() lowerCAmelCase_ : Tuple = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), F"List must be ordered.\n{collection}." lowerCAmelCase_ : Any = int(input('''Enter the number to be found in the list:\n''').strip()) lowerCAmelCase_ : List[str] = ite_ternary_search(collection, target) lowerCAmelCase_ : int = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(F"Iterative search: {target} found at positions: {resulta}") print(F"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
170
1
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : int = inspect.getfile(accelerate.test_utils ) __snake_case : List[Any] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''external_deps''', '''test_metrics.py'''] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 __snake_case : Any = test_metrics @require_cpu def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' debug_launcher(self.test_metrics.main ) @require_single_gpu def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self.test_metrics.main() @require_multi_gpu def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' print(f"""Found {torch.cuda.device_count()} devices.""" ) __snake_case : str = ['''torchrun''', f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(a_ , env=os.environ.copy() )
102
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __A : int = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") require_version("datasets>=1.14.0", "To fix: pip install -r examples/pytorch/audio-classification/requirements.txt") def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : int = 1_6000 ): '''simple docstring''' _UpperCAmelCase = int(round(sample_rate * max_length ) ) if len(_SCREAMING_SNAKE_CASE ) <= sample_length: return wav _UpperCAmelCase = randint(0 , len(_SCREAMING_SNAKE_CASE ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class _a : """simple docstring""" UpperCamelCase__ = field(default=lowerCAmelCase , metadata={"""help""": """Name of a dataset from the datasets package"""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """A file containing the training audio paths and labels."""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """A file containing the validation audio paths and labels."""}) UpperCamelCase__ = field( default="""train""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) UpperCamelCase__ = field( default="""validation""" , metadata={ """help""": ( """The name of the training data set split to use (via the datasets library). Defaults to 'validation'""" ) } , ) UpperCamelCase__ = field( default="""audio""" , metadata={"""help""": """The name of the dataset column containing the audio data. Defaults to 'audio'"""} , ) UpperCamelCase__ = field( default="""label""" , metadata={"""help""": """The name of the dataset column containing the labels. Defaults to 'label'"""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) UpperCamelCase__ = field( default=20 , metadata={"""help""": """Audio clips will be randomly cut to this length during training if the value is set."""} , ) @dataclass class _a : """simple docstring""" UpperCamelCase__ = field( default="""facebook/wav2vec2-base""" , metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} , ) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from the Hub"""}) UpperCamelCase__ = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Name or path of preprocessor config."""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Whether to freeze the feature encoder layers of the model."""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Whether to generate an attention mask in the feature extractor."""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""}) UpperCamelCase__ = field( default=lowerCAmelCase , metadata={"""help""": """Will enable to load a pretrained model whose head dimensions are different."""} , ) def lowercase__ ( self : Optional[Any] )->int: if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( '''The argument `--freeze_feature_extractor` is deprecated and ''' '''will be removed in a future version. Use `--freeze_feature_encoder`''' '''instead. Setting `freeze_feature_encoder==True`.''' , __UpperCamelCase , ) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( '''The argument `--freeze_feature_extractor` is deprecated and ''' '''should not be used in combination with `--freeze_feature_encoder`.''' '''Only make use of `--freeze_feature_encoder`.''' ) def lowercase ( ): '''simple docstring''' _UpperCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_audio_classification''' , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _UpperCAmelCase = training_args.get_process_log_level() logger.setLevel(_SCREAMING_SNAKE_CASE ) transformers.utils.logging.set_verbosity(_SCREAMING_SNAKE_CASE ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} ' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(f'Training/evaluation parameters {training_args}' ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. _UpperCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' '''Use --overwrite_output_dir to train from scratch.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset and prepare it for the audio classification task. _UpperCAmelCase = DatasetDict() _UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( f'--audio_column_name {data_args.audio_column_name} not found in dataset \'{data_args.dataset_name}\'. ' '''Make sure to set `--audio_column_name` to the correct audio column - one of ''' f'{", ".join(raw_datasets["train"].column_names )}.' ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( f'--label_column_name {data_args.label_column_name} not found in dataset \'{data_args.dataset_name}\'. ' '''Make sure to set `--label_column_name` to the correct text column - one of ''' f'{", ".join(raw_datasets["train"].column_names )}.' ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy _UpperCAmelCase = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. _UpperCAmelCase = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) _UpperCAmelCase = feature_extractor.model_input_names[0] def train_transforms(_SCREAMING_SNAKE_CASE : Tuple ): _UpperCAmelCase = [] for audio in batch[data_args.audio_column_name]: _UpperCAmelCase = random_subsample( audio['''array'''] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = feature_extractor(_SCREAMING_SNAKE_CASE , sampling_rate=feature_extractor.sampling_rate ) _UpperCAmelCase = {model_input_name: inputs.get(_SCREAMING_SNAKE_CASE )} _UpperCAmelCase = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(_SCREAMING_SNAKE_CASE : Optional[int] ): _UpperCAmelCase = [audio['''array'''] for audio in batch[data_args.audio_column_name]] _UpperCAmelCase = feature_extractor(_SCREAMING_SNAKE_CASE , sampling_rate=feature_extractor.sampling_rate ) _UpperCAmelCase = {model_input_name: inputs.get(_SCREAMING_SNAKE_CASE )} _UpperCAmelCase = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _UpperCAmelCase = raw_datasets['''train'''].features[data_args.label_column_name].names _UpperCAmelCase , _UpperCAmelCase = {}, {} for i, label in enumerate(_SCREAMING_SNAKE_CASE ): _UpperCAmelCase = str(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = label # Load the accuracy metric from the datasets package _UpperCAmelCase = evaluate.load('''accuracy''' ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(_SCREAMING_SNAKE_CASE : List[str] ): _UpperCAmelCase = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=_SCREAMING_SNAKE_CASE , references=eval_pred.label_ids ) _UpperCAmelCase = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(_SCREAMING_SNAKE_CASE ) , labelaid=_SCREAMING_SNAKE_CASE , idalabel=_SCREAMING_SNAKE_CASE , finetuning_task='''audio-classification''' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _UpperCAmelCase = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=_SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: _UpperCAmelCase = ( raw_datasets['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(_SCREAMING_SNAKE_CASE , output_all_columns=_SCREAMING_SNAKE_CASE ) if training_args.do_eval: if data_args.max_eval_samples is not None: _UpperCAmelCase = ( raw_datasets['''eval'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(_SCREAMING_SNAKE_CASE , output_all_columns=_SCREAMING_SNAKE_CASE ) # Initialize our trainer _UpperCAmelCase = Trainer( model=_SCREAMING_SNAKE_CASE , args=_SCREAMING_SNAKE_CASE , train_dataset=raw_datasets['''train'''] if training_args.do_train else None , eval_dataset=raw_datasets['''eval'''] if training_args.do_eval else None , compute_metrics=_SCREAMING_SNAKE_CASE , tokenizer=_SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: _UpperCAmelCase = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase = last_checkpoint _UpperCAmelCase = trainer.train(resume_from_checkpoint=_SCREAMING_SNAKE_CASE ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _UpperCAmelCase = trainer.evaluate() trainer.log_metrics('''eval''' , _SCREAMING_SNAKE_CASE ) trainer.save_metrics('''eval''' , _SCREAMING_SNAKE_CASE ) # Write model card and (optionally) push to hub _UpperCAmelCase = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''audio-classification''', '''dataset''': data_args.dataset_name, '''tags''': ['''audio-classification'''], } if training_args.push_to_hub: trainer.push_to_hub(**_SCREAMING_SNAKE_CASE ) else: trainer.create_model_card(**_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
260
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a_ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ = StableDiffusionSAGPipeline UpperCAmelCase_ = TEXT_TO_IMAGE_PARAMS UpperCAmelCase_ = TEXT_TO_IMAGE_BATCH_PARAMS UpperCAmelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCAmelCase_ = False def __snake_case ( self : int): '''simple docstring''' torch.manual_seed(0) lowerCAmelCase__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) lowerCAmelCase__ = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowercase__ , set_alpha_to_one=lowercase__ , ) torch.manual_seed(0) lowerCAmelCase__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0) lowerCAmelCase__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) lowerCAmelCase__ = CLIPTextModel(lowercase__) lowerCAmelCase__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') lowerCAmelCase__ = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __snake_case ( self : Dict , lowercase__ : List[Any] , lowercase__ : int=0): '''simple docstring''' if str(lowercase__).startswith('mps'): lowerCAmelCase__ = torch.manual_seed(lowercase__) else: lowerCAmelCase__ = torch.Generator(device=lowercase__).manual_seed(lowercase__) lowerCAmelCase__ = { 'prompt': '.', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 1.0, 'sag_scale': 1.0, 'output_type': 'numpy', } return inputs def __snake_case ( self : Optional[Any]): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class a_ ( unittest.TestCase ): '''simple docstring''' def __snake_case ( self : int): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : List[Any]): '''simple docstring''' lowerCAmelCase__ = StableDiffusionSAGPipeline.from_pretrained('CompVis/stable-diffusion-v1-4') lowerCAmelCase__ = sag_pipe.to(lowercase__) sag_pipe.set_progress_bar_config(disable=lowercase__) lowerCAmelCase__ = '.' lowerCAmelCase__ = torch.manual_seed(0) lowerCAmelCase__ = sag_pipe( [prompt] , generator=lowercase__ , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type='np') lowerCAmelCase__ = output.images lowerCAmelCase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase__ = np.array([0.1_568, 0.1_738, 0.1_695, 0.1_693, 0.1_507, 0.1_705, 0.1_547, 0.1_751, 0.1_949]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5e-2 def __snake_case ( self : Tuple): '''simple docstring''' lowerCAmelCase__ = StableDiffusionSAGPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base') lowerCAmelCase__ = sag_pipe.to(lowercase__) sag_pipe.set_progress_bar_config(disable=lowercase__) lowerCAmelCase__ = '.' lowerCAmelCase__ = torch.manual_seed(0) lowerCAmelCase__ = sag_pipe( [prompt] , generator=lowercase__ , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type='np') lowerCAmelCase__ = output.images lowerCAmelCase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase__ = np.array([0.3_459, 0.2_876, 0.2_537, 0.3_002, 0.2_671, 0.2_160, 0.3_026, 0.2_262, 0.2_371]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5e-2 def __snake_case ( self : Dict): '''simple docstring''' lowerCAmelCase__ = StableDiffusionSAGPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base') lowerCAmelCase__ = sag_pipe.to(lowercase__) sag_pipe.set_progress_bar_config(disable=lowercase__) lowerCAmelCase__ = '.' lowerCAmelCase__ = torch.manual_seed(0) lowerCAmelCase__ = sag_pipe( [prompt] , width=768 , height=512 , generator=lowercase__ , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type='np' , ) lowerCAmelCase__ = output.images assert image.shape == (1, 512, 768, 3)
119
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCAmelCase__ = logging.get_logger(__name__) class a_ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase_ = ['input_values', 'attention_mask'] def __init__( self : Union[str, Any] , lowercase__ : int = 1 , lowercase__ : int = 16_000 , lowercase__ : float = 0.0 , lowercase__ : bool = False , lowercase__ : int = 80 , lowercase__ : int = 16 , lowercase__ : int = 64 , lowercase__ : str = "hann_window" , lowercase__ : float = 1.0 , lowercase__ : float = 80 , lowercase__ : float = 7_600 , lowercase__ : float = 1e-10 , lowercase__ : int = 2 , lowercase__ : bool = True , **lowercase__ : Optional[int] , ): '''simple docstring''' super().__init__(feature_size=lowercase__ , sampling_rate=lowercase__ , padding_value=lowercase__ , **lowercase__) lowerCAmelCase__ = do_normalize lowerCAmelCase__ = return_attention_mask lowerCAmelCase__ = num_mel_bins lowerCAmelCase__ = hop_length lowerCAmelCase__ = win_length lowerCAmelCase__ = win_function lowerCAmelCase__ = frame_signal_scale lowerCAmelCase__ = fmin lowerCAmelCase__ = fmax lowerCAmelCase__ = mel_floor lowerCAmelCase__ = reduction_factor lowerCAmelCase__ = win_length * sampling_rate // 1_000 lowerCAmelCase__ = hop_length * sampling_rate // 1_000 lowerCAmelCase__ = optimal_fft_length(self.sample_size) lowerCAmelCase__ = (self.n_fft // 2) + 1 lowerCAmelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=lowercase__) lowerCAmelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm='slaney' , mel_scale='slaney' , ) if frame_signal_scale != 1.0: warnings.warn( 'The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers' , lowercase__ , ) if reduction_factor != 2.0: warnings.warn( 'The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers' , lowercase__ , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def __snake_case ( lowercase__ : List[np.ndarray] , lowercase__ : List[np.ndarray] , lowercase__ : float = 0.0): '''simple docstring''' if attention_mask is not None: lowerCAmelCase__ = np.array(lowercase__ , np.intaa) lowerCAmelCase__ = [] for vector, length in zip(lowercase__ , attention_mask.sum(-1)): lowerCAmelCase__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7) if length < normed_slice.shape[0]: lowerCAmelCase__ = padding_value normed_input_values.append(lowercase__) else: lowerCAmelCase__ = [(x - x.mean()) / np.sqrt(x.var() + 1e-7) for x in input_values] return normed_input_values def __snake_case ( self : List[Any] , lowercase__ : np.ndarray , ): '''simple docstring''' lowerCAmelCase__ = spectrogram( lowercase__ , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel='log10' , ) return log_mel_spec.T def __call__( self : str , lowercase__ : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , lowercase__ : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , lowercase__ : Union[bool, str, PaddingStrategy] = False , lowercase__ : Optional[int] = None , lowercase__ : bool = False , lowercase__ : Optional[int] = None , lowercase__ : Optional[bool] = None , lowercase__ : Optional[Union[str, TensorType]] = None , lowercase__ : Optional[int] = None , **lowercase__ : Union[str, Any] , ): '''simple docstring''' if audio is None and audio_target is None: raise ValueError('You must provide either `audio` or `audio_target` values.') if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" F""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" F""" {self.sampling_rate} and not {sampling_rate}.""") else: logger.warning( 'It is strongly recommended to pass the ``sampling_rate`` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.') if audio is not None: lowerCAmelCase__ = self._process_audio( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , **lowercase__ , ) else: lowerCAmelCase__ = None if audio_target is not None: lowerCAmelCase__ = self._process_audio( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , **lowercase__ , ) if inputs is None: return inputs_target else: lowerCAmelCase__ = inputs_target['input_values'] lowerCAmelCase__ = inputs_target.get('attention_mask') if decoder_attention_mask is not None: lowerCAmelCase__ = decoder_attention_mask return inputs def __snake_case ( self : Tuple , lowercase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , lowercase__ : bool = False , lowercase__ : Union[bool, str, PaddingStrategy] = False , lowercase__ : Optional[int] = None , lowercase__ : bool = False , lowercase__ : Optional[int] = None , lowercase__ : Optional[bool] = None , lowercase__ : Optional[Union[str, TensorType]] = None , **lowercase__ : Optional[Any] , ): '''simple docstring''' lowerCAmelCase__ = isinstance(lowercase__ , np.ndarray) and len(speech.shape) > 1 if is_batched_numpy and len(speech.shape) > 2: raise ValueError(F"""Only mono-channel audio is supported for input to {self}""") lowerCAmelCase__ = is_batched_numpy or ( isinstance(lowercase__ , (list, tuple)) and (isinstance(speech[0] , (np.ndarray, tuple, list))) ) if is_batched: lowerCAmelCase__ = [np.asarray(lowercase__ , dtype=np.floataa) for speech in speech] elif not is_batched and not isinstance(lowercase__ , np.ndarray): lowerCAmelCase__ = np.asarray(lowercase__ , dtype=np.floataa) elif isinstance(lowercase__ , np.ndarray) and speech.dtype is np.dtype(np.floataa): lowerCAmelCase__ = speech.astype(np.floataa) # always return batch if not is_batched: lowerCAmelCase__ = [speech] # needed to make pad() work on spectrogram inputs lowerCAmelCase__ = self.feature_size # convert into correct format for padding if is_target: lowerCAmelCase__ = [self._extract_mel_features(lowercase__) for waveform in speech] lowerCAmelCase__ = BatchFeature({'input_values': features}) lowerCAmelCase__ = self.num_mel_bins else: lowerCAmelCase__ = BatchFeature({'input_values': speech}) lowerCAmelCase__ = self.pad( lowercase__ , padding=lowercase__ , max_length=lowercase__ , truncation=lowercase__ , pad_to_multiple_of=lowercase__ , return_attention_mask=lowercase__ , **lowercase__ , ) lowerCAmelCase__ = feature_size_hack # convert input values to correct format lowerCAmelCase__ = padded_inputs['input_values'] if not isinstance(input_values[0] , np.ndarray): lowerCAmelCase__ = [np.asarray(lowercase__ , dtype=np.floataa) for array in input_values] elif ( not isinstance(lowercase__ , np.ndarray) and isinstance(input_values[0] , np.ndarray) and input_values[0].dtype is np.dtype(np.floataa) ): lowerCAmelCase__ = [array.astype(np.floataa) for array in input_values] elif isinstance(lowercase__ , np.ndarray) and input_values.dtype is np.dtype(np.floataa): lowerCAmelCase__ = input_values.astype(np.floataa) # convert attention_mask to correct format lowerCAmelCase__ = padded_inputs.get('attention_mask') if attention_mask is not None: lowerCAmelCase__ = [np.asarray(lowercase__ , dtype=np.intaa) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: lowerCAmelCase__ = ( attention_mask if self._get_padding_strategies(lowercase__ , max_length=lowercase__) is not PaddingStrategy.DO_NOT_PAD else None ) lowerCAmelCase__ = self.zero_mean_unit_var_norm( padded_inputs['input_values'] , attention_mask=lowercase__ , padding_value=self.padding_value) if return_tensors is not None: lowerCAmelCase__ = padded_inputs.convert_to_tensors(lowercase__) return padded_inputs def __snake_case ( self : Any): '''simple docstring''' lowerCAmelCase__ = super().to_dict() # Don't serialize these as they are derived from the other properties. lowerCAmelCase__ = ['window', 'mel_filters', 'sample_size', 'sample_stride', 'n_fft', 'n_freqs'] for name in names: if name in output: del output[name] return output
119
1
'''simple docstring''' # This is the module that test_patching.py uses to test patch_submodule() import os # noqa: this is just for tests import os as renamed_os # noqa: this is just for tests from os import path # noqa: this is just for tests from os import path as renamed_path # noqa: this is just for tests from os.path import join # noqa: this is just for tests from os.path import join as renamed_join # noqa: this is just for tests __UpperCAmelCase =open # noqa: we just need to have a builtin inside this module to test it properly
67
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { '''hustvl/yolos-small''': '''https://huggingface.co/hustvl/yolos-small/resolve/main/config.json''', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class __SCREAMING_SNAKE_CASE ( A__ ): A : Any = 'yolos' def __init__( self , SCREAMING_SNAKE_CASE__=768 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=3072 , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=0.02 , SCREAMING_SNAKE_CASE__=1E-12 , SCREAMING_SNAKE_CASE__=[512, 864] , SCREAMING_SNAKE_CASE__=16 , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=100 , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=1 , SCREAMING_SNAKE_CASE__=5 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=5 , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=0.1 , **SCREAMING_SNAKE_CASE__ , ): super().__init__(**SCREAMING_SNAKE_CASE__ ) lowercase : Union[str, Any] = hidden_size lowercase : int = num_hidden_layers lowercase : str = num_attention_heads lowercase : str = intermediate_size lowercase : Dict = hidden_act lowercase : int = hidden_dropout_prob lowercase : Optional[Any] = attention_probs_dropout_prob lowercase : List[Any] = initializer_range lowercase : Optional[int] = layer_norm_eps lowercase : str = image_size lowercase : Dict = patch_size lowercase : str = num_channels lowercase : Optional[int] = qkv_bias lowercase : List[str] = num_detection_tokens lowercase : List[str] = use_mid_position_embeddings lowercase : Dict = auxiliary_loss # Hungarian matcher lowercase : Optional[Any] = class_cost lowercase : Any = bbox_cost lowercase : int = giou_cost # Loss coefficients lowercase : Dict = bbox_loss_coefficient lowercase : Optional[Any] = giou_loss_coefficient lowercase : Tuple = eos_coefficient class __SCREAMING_SNAKE_CASE ( A__ ): A : List[str] = version.parse('1.11' ) @property def __lowerCamelCase ( self ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def __lowerCamelCase ( self ): return 1E-4 @property def __lowerCamelCase ( self ): return 12
337
0
"""simple docstring""" from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract snake_case__ : Optional[Any] = logging.get_logger(__name__) def _snake_case ( _snake_case : int , _snake_case : List[Any] , _snake_case : List[Any] ): return [ int(1000 * (box[0] / width) ), int(1000 * (box[1] / height) ), int(1000 * (box[2] / width) ), int(1000 * (box[3] / height) ), ] def _snake_case ( _snake_case : np.ndarray , _snake_case : Optional[str] , _snake_case : Optional[str] ): lowerCAmelCase : str = to_pil_image(_snake_case ) lowerCAmelCase, lowerCAmelCase : Any = pil_image.size lowerCAmelCase : Optional[Any] = pytesseract.image_to_data(_snake_case , lang=_snake_case , output_type='''dict''' , config=_snake_case ) lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase, lowerCAmelCase : Optional[Any] = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates lowerCAmelCase : int = [idx for idx, word in enumerate(_snake_case ) if not word.strip()] lowerCAmelCase : int = [word for idx, word in enumerate(_snake_case ) if idx not in irrelevant_indices] lowerCAmelCase : Any = [coord for idx, coord in enumerate(_snake_case ) if idx not in irrelevant_indices] lowerCAmelCase : List[str] = [coord for idx, coord in enumerate(_snake_case ) if idx not in irrelevant_indices] lowerCAmelCase : str = [coord for idx, coord in enumerate(_snake_case ) if idx not in irrelevant_indices] lowerCAmelCase : List[str] = [coord for idx, coord in enumerate(_snake_case ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format lowerCAmelCase : List[str] = [] for x, y, w, h in zip(_snake_case , _snake_case , _snake_case , _snake_case ): lowerCAmelCase : Tuple = [x, y, x + w, y + h] actual_boxes.append(_snake_case ) # finally, normalize the bounding boxes lowerCAmelCase : List[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(_snake_case , _snake_case , _snake_case ) ) assert len(_snake_case ) == len(_snake_case ), "Not as many words as there are bounding boxes" return words, normalized_boxes class snake_case_( a__ ): __UpperCamelCase = ['''pixel_values'''] def __init__( self : Union[str, Any] , UpperCamelCase_ : bool = True , UpperCamelCase_ : Dict[str, int] = None , UpperCamelCase_ : PILImageResampling = PILImageResampling.BILINEAR , UpperCamelCase_ : bool = True , UpperCamelCase_ : float = 1 / 2_5_5 , UpperCamelCase_ : bool = True , UpperCamelCase_ : Union[float, Iterable[float]] = None , UpperCamelCase_ : Union[float, Iterable[float]] = None , UpperCamelCase_ : bool = True , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = "" , **UpperCamelCase_ : int , ): super().__init__(**UpperCamelCase_ ) lowerCAmelCase : Any = size if size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} lowerCAmelCase : Optional[Any] = get_size_dict(UpperCamelCase_ ) lowerCAmelCase : Optional[Any] = do_resize lowerCAmelCase : str = size lowerCAmelCase : List[str] = resample lowerCAmelCase : Optional[int] = do_rescale lowerCAmelCase : int = rescale_value lowerCAmelCase : Optional[Any] = do_normalize lowerCAmelCase : str = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCAmelCase : str = image_std if image_std is not None else IMAGENET_STANDARD_STD lowerCAmelCase : Any = apply_ocr lowerCAmelCase : Tuple = ocr_lang lowerCAmelCase : Union[str, Any] = tesseract_config def lowerCamelCase__ ( self : Optional[Any] , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : Dict[str, int] , UpperCamelCase_ : PILImageResampling = PILImageResampling.BILINEAR , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : str , ): lowerCAmelCase : List[str] = get_size_dict(UpperCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) lowerCAmelCase : int = (size['''height'''], size['''width''']) return resize(UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : Optional[int] , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : Union[int, float] , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : Tuple , ): return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : int , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : Union[float, Iterable[float]] , UpperCamelCase_ : Union[float, Iterable[float]] , UpperCamelCase_ : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_ : Any , ): return normalize(UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase__ ( self : int , UpperCamelCase_ : ImageInput , UpperCamelCase_ : bool = None , UpperCamelCase_ : Dict[str, int] = None , UpperCamelCase_ : List[Any]=None , UpperCamelCase_ : bool = None , UpperCamelCase_ : float = None , UpperCamelCase_ : bool = None , UpperCamelCase_ : Union[float, Iterable[float]] = None , UpperCamelCase_ : Union[float, Iterable[float]] = None , UpperCamelCase_ : bool = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , UpperCamelCase_ : ChannelDimension = ChannelDimension.FIRST , **UpperCamelCase_ : Union[str, Any] , ): lowerCAmelCase : Union[str, Any] = do_resize if do_resize is not None else self.do_resize lowerCAmelCase : str = size if size is not None else self.size lowerCAmelCase : Optional[Any] = get_size_dict(UpperCamelCase_ ) lowerCAmelCase : Any = resample if resample is not None else self.resample lowerCAmelCase : Dict = do_rescale if do_rescale is not None else self.do_rescale lowerCAmelCase : str = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCAmelCase : List[str] = do_normalize if do_normalize is not None else self.do_normalize lowerCAmelCase : Dict = image_mean if image_mean is not None else self.image_mean lowerCAmelCase : int = image_std if image_std is not None else self.image_std lowerCAmelCase : int = apply_ocr if apply_ocr is not None else self.apply_ocr lowerCAmelCase : Optional[int] = ocr_lang if ocr_lang is not None else self.ocr_lang lowerCAmelCase : Optional[Any] = tesseract_config if tesseract_config is not None else self.tesseract_config lowerCAmelCase : List[Any] = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''If do_normalize is True, image_mean and image_std must be specified.''' ) # All transformations expect numpy arrays. lowerCAmelCase : Optional[Any] = [to_numpy_array(UpperCamelCase_ ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , '''pytesseract''' ) lowerCAmelCase : Optional[int] = [] lowerCAmelCase : str = [] for image in images: lowerCAmelCase, lowerCAmelCase : Union[str, Any] = apply_tesseract(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) words_batch.append(UpperCamelCase_ ) boxes_batch.append(UpperCamelCase_ ) if do_resize: lowerCAmelCase : int = [self.resize(image=UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ ) for image in images] if do_rescale: lowerCAmelCase : Dict = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_normalize: lowerCAmelCase : int = [self.normalize(image=UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ ) for image in images] lowerCAmelCase : str = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] lowerCAmelCase : Optional[Any] = BatchFeature(data={'''pixel_values''': images} , tensor_type=UpperCamelCase_ ) if apply_ocr: lowerCAmelCase : Tuple = words_batch lowerCAmelCase : Dict = boxes_batch return data
314
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer snake_case__ : str = logging.get_logger(__name__) snake_case__ : List[str] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} snake_case__ : str = { '''vocab_file''': { '''bert-base-uncased''': '''https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt''', '''bert-large-uncased''': '''https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt''', '''bert-base-cased''': '''https://huggingface.co/bert-base-cased/resolve/main/vocab.txt''', '''bert-large-cased''': '''https://huggingface.co/bert-large-cased/resolve/main/vocab.txt''', '''bert-base-multilingual-uncased''': ( '''https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt''' ), '''bert-base-multilingual-cased''': '''https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt''', '''bert-base-chinese''': '''https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt''', '''bert-base-german-cased''': '''https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt''', '''bert-large-uncased-whole-word-masking''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt''' ), '''bert-large-cased-whole-word-masking''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt''' ), '''bert-large-uncased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt''' ), '''bert-large-cased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt''' ), '''bert-base-cased-finetuned-mrpc''': ( '''https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt''' ), '''bert-base-german-dbmdz-cased''': '''https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt''', '''bert-base-german-dbmdz-uncased''': ( '''https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt''' ), '''TurkuNLP/bert-base-finnish-cased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt''' ), '''TurkuNLP/bert-base-finnish-uncased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt''' ), '''wietsedv/bert-base-dutch-cased''': ( '''https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''bert-base-uncased''': '''https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json''', '''bert-large-uncased''': '''https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json''', '''bert-base-cased''': '''https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json''', '''bert-large-cased''': '''https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json''', '''bert-base-multilingual-uncased''': ( '''https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json''' ), '''bert-base-multilingual-cased''': ( '''https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json''' ), '''bert-base-chinese''': '''https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json''', '''bert-base-german-cased''': '''https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json''', '''bert-large-uncased-whole-word-masking''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json''' ), '''bert-large-cased-whole-word-masking''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json''' ), '''bert-large-uncased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json''' ), '''bert-large-cased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json''' ), '''bert-base-cased-finetuned-mrpc''': ( '''https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json''' ), '''bert-base-german-dbmdz-cased''': ( '''https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json''' ), '''bert-base-german-dbmdz-uncased''': ( '''https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json''' ), '''TurkuNLP/bert-base-finnish-cased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json''' ), '''TurkuNLP/bert-base-finnish-uncased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json''' ), '''wietsedv/bert-base-dutch-cased''': ( '''https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json''' ), }, } snake_case__ : Union[str, Any] = { '''bert-base-uncased''': 512, '''bert-large-uncased''': 512, '''bert-base-cased''': 512, '''bert-large-cased''': 512, '''bert-base-multilingual-uncased''': 512, '''bert-base-multilingual-cased''': 512, '''bert-base-chinese''': 512, '''bert-base-german-cased''': 512, '''bert-large-uncased-whole-word-masking''': 512, '''bert-large-cased-whole-word-masking''': 512, '''bert-large-uncased-whole-word-masking-finetuned-squad''': 512, '''bert-large-cased-whole-word-masking-finetuned-squad''': 512, '''bert-base-cased-finetuned-mrpc''': 512, '''bert-base-german-dbmdz-cased''': 512, '''bert-base-german-dbmdz-uncased''': 512, '''TurkuNLP/bert-base-finnish-cased-v1''': 512, '''TurkuNLP/bert-base-finnish-uncased-v1''': 512, '''wietsedv/bert-base-dutch-cased''': 512, } snake_case__ : Optional[Any] = { '''bert-base-uncased''': {'''do_lower_case''': True}, '''bert-large-uncased''': {'''do_lower_case''': True}, '''bert-base-cased''': {'''do_lower_case''': False}, '''bert-large-cased''': {'''do_lower_case''': False}, '''bert-base-multilingual-uncased''': {'''do_lower_case''': True}, '''bert-base-multilingual-cased''': {'''do_lower_case''': False}, '''bert-base-chinese''': {'''do_lower_case''': False}, '''bert-base-german-cased''': {'''do_lower_case''': False}, '''bert-large-uncased-whole-word-masking''': {'''do_lower_case''': True}, '''bert-large-cased-whole-word-masking''': {'''do_lower_case''': False}, '''bert-large-uncased-whole-word-masking-finetuned-squad''': {'''do_lower_case''': True}, '''bert-large-cased-whole-word-masking-finetuned-squad''': {'''do_lower_case''': False}, '''bert-base-cased-finetuned-mrpc''': {'''do_lower_case''': False}, '''bert-base-german-dbmdz-cased''': {'''do_lower_case''': False}, '''bert-base-german-dbmdz-uncased''': {'''do_lower_case''': True}, '''TurkuNLP/bert-base-finnish-cased-v1''': {'''do_lower_case''': False}, '''TurkuNLP/bert-base-finnish-uncased-v1''': {'''do_lower_case''': True}, '''wietsedv/bert-base-dutch-cased''': {'''do_lower_case''': False}, } class snake_case_( a__ ): __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_INIT_CONFIGURATION __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = BertTokenizer def __init__( self : int , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : str=True , UpperCamelCase_ : Dict="[UNK]" , UpperCamelCase_ : Any="[SEP]" , UpperCamelCase_ : Any="[PAD]" , UpperCamelCase_ : Tuple="[CLS]" , UpperCamelCase_ : List[Any]="[MASK]" , UpperCamelCase_ : Optional[Any]=True , UpperCamelCase_ : Tuple=None , **UpperCamelCase_ : Optional[int] , ): super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , do_lower_case=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , tokenize_chinese_chars=UpperCamelCase_ , strip_accents=UpperCamelCase_ , **UpperCamelCase_ , ) lowerCAmelCase : Any = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , UpperCamelCase_ ) != do_lower_case or normalizer_state.get('''strip_accents''' , UpperCamelCase_ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , UpperCamelCase_ ) != tokenize_chinese_chars ): lowerCAmelCase : Optional[int] = getattr(UpperCamelCase_ , normalizer_state.pop('''type''' ) ) lowerCAmelCase : Tuple = do_lower_case lowerCAmelCase : Union[str, Any] = strip_accents lowerCAmelCase : Tuple = tokenize_chinese_chars lowerCAmelCase : str = normalizer_class(**UpperCamelCase_ ) lowerCAmelCase : Optional[int] = do_lower_case def lowerCamelCase__ ( self : List[str] , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple=None ): lowerCAmelCase : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase__ ( self : List[Any] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None ): lowerCAmelCase : Optional[Any] = [self.sep_token_id] lowerCAmelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self : Dict , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None ): lowerCAmelCase : str = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_ ) return tuple(UpperCamelCase_ )
314
1
A__ = '''Alexander Joslin''' import operator as op from .stack import Stack def _lowerCAmelCase ( __lowerCAmelCase ) -> int: """simple docstring""" snake_case__ : Union[str, Any] = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} snake_case__ : Stack[int] = Stack() snake_case__ : Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__lowerCAmelCase ) ) elif i in operators: # RULE 2 operator_stack.push(__lowerCAmelCase ) elif i == ")": # RULE 4 snake_case__ : Any = operator_stack.peek() operator_stack.pop() snake_case__ : Union[str, Any] = operand_stack.peek() operand_stack.pop() snake_case__ : List[str] = operand_stack.peek() operand_stack.pop() snake_case__ : Optional[Any] = operators[opr](__lowerCAmelCase , __lowerCAmelCase ) operand_stack.push(__lowerCAmelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": A__ = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(f"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
230
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets A__ = datasets.logging.get_logger(__name__) A__ = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' A__ = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' A__ = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=True , __lowerCAmelCase=False , __lowerCAmelCase="dummy_doc" ) -> int: """simple docstring""" snake_case__ : Dict = {doc: key_lines} snake_case__ : Any = {doc: sys_lines} snake_case__ : Dict = {} snake_case__ : List[str] = 0 snake_case__ : Optional[Any] = 0 snake_case__ : Optional[Any] = 0 snake_case__ : Dict = 0 snake_case__ : List[Any] = 0 snake_case__ : List[Any] = 0 snake_case__ , snake_case__ : Tuple = reader.get_doc_mentions(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: snake_case__ : str = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ , snake_case__ : int = reader.get_doc_mentions(__lowerCAmelCase , sys_doc_lines[doc] , __lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: snake_case__ : Union[str, Any] = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) if remove_nested: snake_case__ , snake_case__ : Dict = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters snake_case__ , snake_case__ : Optional[int] = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters snake_case__ : Any = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : Optional[int] = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : List[Any] = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( '''Number of resulting singleton clusters in the key ''' f"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( f"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ '''files, respectively''' ) return doc_coref_infos def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: """simple docstring""" snake_case__ : Optional[Any] = get_coref_infos(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ : str = {} snake_case__ : Optional[int] = 0 snake_case__ : List[Any] = 0 for name, metric in metrics: snake_case__ , snake_case__ , snake_case__ : Any = evaluator.evaluate_documents(__lowerCAmelCase , __lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"""{name}/recall""": recall, f"""{name}/precision""": precision, f"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , f"""Recall: {recall * 100:.2f}""" , f""" Precision: {precision * 100:.2f}""" , f""" F1: {fa * 100:.2f}""" , ) if conll_subparts_num == 3: snake_case__ : int = (conll / 3) * 100 logger.info(f"""CoNLL score: {conll:.2f}""" ) output_scores.update({'''conll_score''': conll} ) return output_scores def _lowerCAmelCase ( __lowerCAmelCase ) -> List[str]: """simple docstring""" snake_case__ : str = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: snake_case__ : List[Any] = line.split()[5] if not parse_col == "-": snake_case__ : Optional[Any] = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): def __lowerCamelCase ( self :Dict ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) ,codebase_urls=['''https://github.com/ns-moosavi/coval'''] ,reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] ,) def __lowerCamelCase ( self :Any ,__lowercase :List[Any] ,__lowercase :int ,__lowercase :str=True ,__lowercase :Optional[int]=False ,__lowercase :Optional[Any]=False ,__lowercase :Tuple=False ): snake_case__ : Optional[Any] = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: snake_case__ : Optional[int] = util.check_gold_parse_annotation(__lowercase ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" snake_case__ : Any = evaluate( key_lines=__lowercase ,sys_lines=__lowercase ,metrics=__lowercase ,NP_only=__lowercase ,remove_nested=__lowercase ,keep_singletons=__lowercase ,min_span=__lowercase ,) return score
230
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A : Tuple = logging.get_logger(__name__) A : List[Any] = { "uclanlp/visualbert-vqa": "https://huggingface.co/uclanlp/visualbert-vqa/resolve/main/config.json", "uclanlp/visualbert-vqa-pre": "https://huggingface.co/uclanlp/visualbert-vqa-pre/resolve/main/config.json", "uclanlp/visualbert-vqa-coco-pre": ( "https://huggingface.co/uclanlp/visualbert-vqa-coco-pre/resolve/main/config.json" ), "uclanlp/visualbert-vcr": "https://huggingface.co/uclanlp/visualbert-vcr/resolve/main/config.json", "uclanlp/visualbert-vcr-pre": "https://huggingface.co/uclanlp/visualbert-vcr-pre/resolve/main/config.json", "uclanlp/visualbert-vcr-coco-pre": ( "https://huggingface.co/uclanlp/visualbert-vcr-coco-pre/resolve/main/config.json" ), "uclanlp/visualbert-nlvr2": "https://huggingface.co/uclanlp/visualbert-nlvr2/resolve/main/config.json", "uclanlp/visualbert-nlvr2-pre": "https://huggingface.co/uclanlp/visualbert-nlvr2-pre/resolve/main/config.json", "uclanlp/visualbert-nlvr2-coco-pre": ( "https://huggingface.co/uclanlp/visualbert-nlvr2-coco-pre/resolve/main/config.json" ) # See all VisualBERT models at https://huggingface.co/models?filter=visual_bert } class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : Any ="""visual_bert""" def __init__( self , __a=3_05_22 , __a=7_68 , __a=5_12 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=2 , __a=0.0_2 , __a=1e-1_2 , __a=False , __a=True , __a=1 , __a=0 , __a=2 , **__a , ): super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = hidden_size __lowerCAmelCase = visual_embedding_dim __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = initializer_range __lowerCAmelCase = type_vocab_size __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = bypass_transformer __lowerCAmelCase = special_visual_initialize
259
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : List[str] = { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/config.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/config.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json" ), "distilbert-base-uncased-finetuned-sst-2-english": ( "https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json" ), } class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : List[str] ="""distilbert""" __UpperCAmelCase : int ={ """hidden_size""": """dim""", """num_attention_heads""": """n_heads""", """num_hidden_layers""": """n_layers""", } def __init__( self , __a=3_05_22 , __a=5_12 , __a=False , __a=6 , __a=12 , __a=7_68 , __a=4 * 7_68 , __a=0.1 , __a=0.1 , __a="gelu" , __a=0.0_2 , __a=0.1 , __a=0.2 , __a=0 , **__a , ): __lowerCAmelCase = vocab_size __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = sinusoidal_pos_embds __lowerCAmelCase = n_layers __lowerCAmelCase = n_heads __lowerCAmelCase = dim __lowerCAmelCase = hidden_dim __lowerCAmelCase = dropout __lowerCAmelCase = attention_dropout __lowerCAmelCase = activation __lowerCAmelCase = initializer_range __lowerCAmelCase = qa_dropout __lowerCAmelCase = seq_classif_dropout super().__init__(**__a , pad_token_id=__a ) class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' @property def snake_case ( self ): if self.task == "multiple-choice": __lowerCAmelCase = {0: "batch", 1: "choice", 2: "sequence"} else: __lowerCAmelCase = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
259
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class __A ( unittest.TestCase ): def __init__( self , a__ , a__=7 , a__=3 , a__=30 , a__=400 , a__=True , a__=None , a__=0.9 , a__=None , a__=True , a__=[0.5, 0.5, 0.5] , a__=[0.5, 0.5, 0.5] , ): _lowerCAmelCase : Dict = size if size is not None else {"""shortest_edge""": 30} _lowerCAmelCase : Union[str, Any] = crop_size if crop_size is not None else {"""height""": 30, """width""": 30} _lowerCAmelCase : List[str] = parent _lowerCAmelCase : int = batch_size _lowerCAmelCase : Optional[Any] = num_channels _lowerCAmelCase : List[Any] = min_resolution _lowerCAmelCase : Tuple = max_resolution _lowerCAmelCase : Optional[int] = do_resize_and_center_crop _lowerCAmelCase : Dict = size _lowerCAmelCase : int = crop_pct _lowerCAmelCase : Optional[Any] = crop_size _lowerCAmelCase : Tuple = do_normalize _lowerCAmelCase : List[str] = image_mean _lowerCAmelCase : Optional[Any] = image_std def __A ( self ): return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : List[Any] = PoolFormerImageProcessor if is_vision_available() else None def __A ( self ): _lowerCAmelCase : Tuple = PoolFormerImageProcessingTester(self ) @property def __A ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __A ( self ): _lowerCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a__ , """do_resize_and_center_crop""" ) ) self.assertTrue(hasattr(a__ , """size""" ) ) self.assertTrue(hasattr(a__ , """crop_pct""" ) ) self.assertTrue(hasattr(a__ , """do_normalize""" ) ) self.assertTrue(hasattr(a__ , """image_mean""" ) ) self.assertTrue(hasattr(a__ , """image_std""" ) ) def __A ( self ): _lowerCAmelCase : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 30} ) self.assertEqual(image_processor.crop_size , {"""height""": 30, """width""": 30} ) _lowerCAmelCase : str = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def __A ( self ): pass def __A ( self ): # Initialize image_processing _lowerCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=a__ ) for image in image_inputs: self.assertIsInstance(a__ , Image.Image ) # Test not batched input _lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _lowerCAmelCase : str = image_processing(a__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): # Initialize image_processing _lowerCAmelCase : int = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCAmelCase : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=a__ , numpify=a__ ) for image in image_inputs: self.assertIsInstance(a__ , np.ndarray ) # Test not batched input _lowerCAmelCase : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _lowerCAmelCase : List[Any] = image_processing(a__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def __A ( self ): # Initialize image_processing _lowerCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=a__ , torchify=a__ ) for image in image_inputs: self.assertIsInstance(a__ , torch.Tensor ) # Test not batched input _lowerCAmelCase : Dict = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _lowerCAmelCase : Any = image_processing(a__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
44
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCamelCase = { '''configuration_trajectory_transformer''': [ '''TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TrajectoryTransformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TrajectoryTransformerModel''', '''TrajectoryTransformerPreTrainedModel''', '''load_tf_weights_in_trajectory_transformer''', ] if TYPE_CHECKING: from .configuration_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TrajectoryTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trajectory_transformer import ( TRAJECTORY_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TrajectoryTransformerModel, TrajectoryTransformerPreTrainedModel, load_tf_weights_in_trajectory_transformer, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
254
0
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowerCAmelCase : Tuple ={ """facebook/mask2former-swin-small-coco-instance""": ( """https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json""" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } __lowerCAmelCase : Optional[Any] =logging.get_logger(__name__) class _A ( lowerCAmelCase ): snake_case__ : Dict = 'mask2former' snake_case__ : Union[str, Any] = ['swin'] snake_case__ : Any = {'hidden_size': 'hidden_dim'} def __init__( self , __lowerCAmelCase = None , __lowerCAmelCase = 256 , __lowerCAmelCase = 256 , __lowerCAmelCase = 256 , __lowerCAmelCase = 1024 , __lowerCAmelCase = "relu" , __lowerCAmelCase = 6 , __lowerCAmelCase = 10 , __lowerCAmelCase = 8 , __lowerCAmelCase = 0.0 , __lowerCAmelCase = 2048 , __lowerCAmelCase = False , __lowerCAmelCase = False , __lowerCAmelCase = 4 , __lowerCAmelCase = 255 , __lowerCAmelCase = 100 , __lowerCAmelCase = 0.1 , __lowerCAmelCase = 2.0 , __lowerCAmelCase = 5.0 , __lowerCAmelCase = 5.0 , __lowerCAmelCase = 1_2544 , __lowerCAmelCase = 3.0 , __lowerCAmelCase = 0.7_5 , __lowerCAmelCase = 0.0_2 , __lowerCAmelCase = 1.0 , __lowerCAmelCase = True , __lowerCAmelCase = [4, 8, 16, 32] , __lowerCAmelCase = None , **__lowerCAmelCase , ): """simple docstring""" if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.""" ) lowercase = CONFIG_MAPPING["""swin"""]( image_size=224 , in_channels=3 , patch_size=4 , embed_dim=96 , depths=[2, 2, 18, 2] , num_heads=[3, 6, 12, 24] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=__lowerCAmelCase , out_features=["""stage1""", """stage2""", """stage3""", """stage4"""] , ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ): lowercase = backbone_config.pop("""model_type""" ) lowercase = CONFIG_MAPPING[backbone_model_type] lowercase = config_class.from_dict(__lowerCAmelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f'Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. ' f'Supported model types: {",".join(self.backbones_supported )}' ) lowercase = backbone_config lowercase = feature_size lowercase = mask_feature_size lowercase = hidden_dim lowercase = encoder_feedforward_dim lowercase = activation_function lowercase = encoder_layers lowercase = decoder_layers lowercase = num_attention_heads lowercase = dropout lowercase = dim_feedforward lowercase = pre_norm lowercase = enforce_input_projection lowercase = common_stride lowercase = ignore_value lowercase = num_queries lowercase = no_object_weight lowercase = class_weight lowercase = mask_weight lowercase = dice_weight lowercase = train_num_points lowercase = oversample_ratio lowercase = importance_sample_ratio lowercase = init_std lowercase = init_xavier_std lowercase = use_auxiliary_loss lowercase = feature_strides lowercase = output_auxiliary_logits lowercase = decoder_layers super().__init__(**__lowerCAmelCase ) @classmethod def A__ ( cls , __lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" return cls( backbone_config=__lowerCAmelCase , **__lowerCAmelCase , ) def A__ ( self ): """simple docstring""" lowercase = copy.deepcopy(self.__dict__ ) lowercase = self.backbone_config.to_dict() lowercase = self.__class__.model_type return output
371
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING __lowerCAmelCase : Union[str, Any] =logging.get_logger(__name__) @add_end_docstrings(lowerCAmelCase ) class _A ( lowerCAmelCase ): def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" super().__init__(*__lowerCAmelCase , **__lowerCAmelCase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def A__ ( self , __lowerCAmelCase=None ): """simple docstring""" lowercase = {} if top_k is not None: lowercase = top_k return {}, {}, postprocess_params def __call__( self , __lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" return super().__call__(__lowerCAmelCase , **__lowerCAmelCase ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" lowercase = load_image(__lowerCAmelCase ) lowercase = self.image_processor(images=__lowerCAmelCase , return_tensors=self.framework ) return model_inputs def A__ ( self , __lowerCAmelCase ): """simple docstring""" lowercase = self.model(**__lowerCAmelCase ) return model_outputs def A__ ( self , __lowerCAmelCase , __lowerCAmelCase=5 ): """simple docstring""" if top_k > self.model.config.num_labels: lowercase = self.model.config.num_labels if self.framework == "pt": lowercase = model_outputs.logits.softmax(-1 )[0] lowercase , lowercase = probs.topk(__lowerCAmelCase ) elif self.framework == "tf": lowercase = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase = tf.math.top_k(__lowerCAmelCase , k=__lowerCAmelCase ) lowercase , lowercase = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'Unsupported framework: {self.framework}' ) lowercase = scores.tolist() lowercase = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(__lowerCAmelCase , __lowerCAmelCase )]
32
0
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() _UpperCamelCase : Optional[Any] = logging.get_logger(__name__) _UpperCamelCase : List[Any] = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'adapter_layer': 'encoder.layers.*.adapter_layer', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', 'pooling_layer.linear': 'projector', 'pooling_layer.projection': 'classifier', } _UpperCamelCase : List[Any] = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', 'projector', 'classifier', ] def __UpperCAmelCase ( A : Optional[Any] ) -> Optional[Any]: UpperCAmelCase_ : List[Any] = {} with open(A , '''r''' ) as file: for line_number, line in enumerate(A ): UpperCAmelCase_ : Any = line.strip() if line: UpperCAmelCase_ : Optional[int] = line.split() UpperCAmelCase_ : List[Any] = line_number UpperCAmelCase_ : str = words[0] UpperCAmelCase_ : Dict = value return result def __UpperCAmelCase ( A : str , A : Any , A : Dict , A : int , A : Optional[Any] ) -> str: for attribute in key.split('''.''' ): UpperCAmelCase_ : Union[str, Any] = getattr(A , A ) UpperCAmelCase_ : Optional[int] = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(A ): UpperCAmelCase_ : Optional[int] = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase_ : str = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase_ : Optional[Any] = getattr(A , A ).shape elif weight_type is not None and weight_type == "param": UpperCAmelCase_ : int = hf_pointer for attribute in hf_param_name.split('''.''' ): UpperCAmelCase_ : List[Any] = getattr(A , A ) UpperCAmelCase_ : Dict = shape_pointer.shape # let's reduce dimension UpperCAmelCase_ : Optional[int] = value[0] else: UpperCAmelCase_ : int = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": UpperCAmelCase_ : Optional[int] = value elif weight_type == "weight_g": UpperCAmelCase_ : int = value elif weight_type == "weight_v": UpperCAmelCase_ : Union[str, Any] = value elif weight_type == "bias": UpperCAmelCase_ : Optional[int] = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): UpperCAmelCase_ : int = getattr(A , A ) UpperCAmelCase_ : str = value else: UpperCAmelCase_ : Any = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def __UpperCAmelCase ( A : Tuple , A : Dict , A : List[str] , A : int , A : List[str] ) -> Union[str, Any]: UpperCAmelCase_ : Union[str, Any] = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(A ): UpperCAmelCase_ : str = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase_ : str = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase_ : int = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": UpperCAmelCase_ : Union[str, Any] = '''.'''.join([key, hf_param_name] ) else: UpperCAmelCase_ : Optional[int] = key UpperCAmelCase_ : str = value if '''lm_head''' in full_key else value[0] _UpperCamelCase : int = { 'W_a': 'linear_1.weight', 'W_b': 'linear_2.weight', 'b_a': 'linear_1.bias', 'b_b': 'linear_2.bias', 'ln_W': 'norm.weight', 'ln_b': 'norm.bias', } def __UpperCAmelCase ( A : Any , A : Dict , A : Any=None , A : Tuple=None ) -> int: UpperCAmelCase_ : Optional[Any] = False for key, mapped_key in MAPPING.items(): UpperCAmelCase_ : Dict = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: UpperCAmelCase_ : Any = True if "*" in mapped_key: UpperCAmelCase_ : Optional[int] = name.split(A )[0].split('''.''' )[-2] UpperCAmelCase_ : Optional[Any] = mapped_key.replace('''*''' , A ) if "weight_g" in name: UpperCAmelCase_ : List[Any] = '''weight_g''' elif "weight_v" in name: UpperCAmelCase_ : List[str] = '''weight_v''' elif "bias" in name: UpperCAmelCase_ : List[str] = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase_ : int = '''weight''' else: UpperCAmelCase_ : Union[str, Any] = None if hf_dict is not None: rename_dict(A , A , A , A , A ) else: set_recursively(A , A , A , A , A ) return is_used return is_used def __UpperCAmelCase ( A : Tuple , A : List[Any] , A : int ) -> Tuple: UpperCAmelCase_ : Any = [] UpperCAmelCase_ : int = fairseq_model.state_dict() UpperCAmelCase_ : Any = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase_ : Any = False if "conv_layers" in name: load_conv_layer( A , A , A , A , hf_model.config.feat_extract_norm == '''group''' , ) UpperCAmelCase_ : Optional[int] = True else: UpperCAmelCase_ : Tuple = load_wavaveca_layer(A , A , A ) if not is_used: unused_weights.append(A ) logger.warning(F"Unused weights: {unused_weights}" ) def __UpperCAmelCase ( A : Tuple , A : int , A : int , A : int , A : Tuple ) -> str: UpperCAmelCase_ : Any = full_name.split('''conv_layers.''' )[-1] UpperCAmelCase_ : str = name.split('''.''' ) UpperCAmelCase_ : int = int(items[0] ) UpperCAmelCase_ : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) UpperCAmelCase_ : List[str] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) UpperCAmelCase_ : str = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found." ) UpperCAmelCase_ : Tuple = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found." ) UpperCAmelCase_ : Optional[int] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(A ) @torch.no_grad() def __UpperCAmelCase ( A : List[Any] , A : Tuple , A : Tuple=None , A : Tuple=None , A : Dict=True , A : Union[str, Any]=False ) -> Dict: if config_path is not None: UpperCAmelCase_ : Dict = WavaVecaConfig.from_pretrained(A ) else: UpperCAmelCase_ : List[str] = WavaVecaConfig() if is_seq_class: UpperCAmelCase_ : List[Any] = read_txt_into_dict(A ) UpperCAmelCase_ : List[str] = idalabel UpperCAmelCase_ : List[str] = WavaVecaForSequenceClassification(A ) UpperCAmelCase_ : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=A , return_attention_mask=A , ) feature_extractor.save_pretrained(A ) elif is_finetuned: if dict_path: UpperCAmelCase_ : Dict = Dictionary.load(A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase_ : int = target_dict.pad_index UpperCAmelCase_ : Tuple = target_dict.bos_index UpperCAmelCase_ : Tuple = target_dict.eos_index UpperCAmelCase_ : Dict = len(target_dict.symbols ) UpperCAmelCase_ : Any = os.path.join(A , '''vocab.json''' ) if not os.path.isdir(A ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(A ) ) return os.makedirs(A , exist_ok=A ) UpperCAmelCase_ : Tuple = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase_ : int = 0 UpperCAmelCase_ : List[str] = 1 with open(A , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(A , A ) UpperCAmelCase_ : Union[str, Any] = WavaVecaCTCTokenizer( A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=A , ) UpperCAmelCase_ : Tuple = True if config.feat_extract_norm == '''layer''' else False UpperCAmelCase_ : Union[str, Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=A , return_attention_mask=A , ) UpperCAmelCase_ : List[Any] = WavaVecaProcessor(feature_extractor=A , tokenizer=A ) processor.save_pretrained(A ) UpperCAmelCase_ : Dict = WavaVecaForCTC(A ) else: UpperCAmelCase_ : Tuple = WavaVecaForPreTraining(A ) if is_finetuned or is_seq_class: UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: UpperCAmelCase_ : Dict = argparse.Namespace(task='''audio_pretraining''' ) UpperCAmelCase_ : List[str] = fairseq.tasks.setup_task(A ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : int = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=A ) UpperCAmelCase_ : Optional[int] = model[0].eval() recursively_load_weights(A , A , not is_finetuned ) hf_wavavec.save_pretrained(A ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--not_finetuned', action='store_true', help='Whether the model to convert is a fine-tuned model or not' ) parser.add_argument( '--is_seq_class', action='store_true', help='Whether the model to convert is a fine-tuned sequence classification model or not', ) _UpperCamelCase : str = parser.parse_args() _UpperCamelCase : List[Any] = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
304
'''simple docstring''' import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Union[str, Any] = logging.getLogger(__name__) _UpperCamelCase : Optional[int] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : str = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class snake_case__ : a_ = field( default=UpperCamelCase , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) a_ = field( default=UpperCamelCase , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(UpperCamelCase)} , ) a_ = field( default=UpperCamelCase , metadata={"help": "Pretrained config name or path if not the same as model_name"}) a_ = field( default=UpperCamelCase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) a_ = field( default=UpperCamelCase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class snake_case__ : a_ = field( default=UpperCamelCase , metadata={"help": "The input training data file (a text file)."}) a_ = field( default=UpperCamelCase , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) a_ = field( default=UpperCamelCase , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) a_ = field( default=UpperCamelCase , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) a_ = field(default=UpperCamelCase , metadata={"help": "Whether ot not to use whole word mask."}) a_ = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) a_ = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) a_ = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) a_ = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) a_ = field( default=UpperCamelCase , metadata={"help": "Overwrite the cached training and evaluation sets"}) def __UpperCAmelCase ( A : DataTrainingArguments , A : PreTrainedTokenizer , A : bool = False , A : Optional[str] = None , ) -> List[Any]: def _dataset(A : Dict , A : str=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('''You need to set world whole masking and mlm to True for Chinese Whole Word Mask''' ) return LineByLineWithRefDataset( tokenizer=A , file_path=A , block_size=args.block_size , ref_path=A , ) return LineByLineTextDataset(tokenizer=A , file_path=A , block_size=args.block_size ) else: return TextDataset( tokenizer=A , file_path=A , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=A , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(A ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def __UpperCAmelCase ( ) -> Optional[Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. UpperCAmelCase_ : str = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : str = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( '''Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ''' '''or remove the --do_eval argument.''' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( '''Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s''' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('''Training/evaluation parameters %s''' , A ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: UpperCAmelCase_ : List[str] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCAmelCase_ : List[str] = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: UpperCAmelCase_ : List[Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.tokenizer_name: UpperCAmelCase_ : str = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: UpperCAmelCase_ : List[str] = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( '''You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another''' ''' script, save it,and load it from here, using --tokenizer_name''' ) if model_args.model_name_or_path: UpperCAmelCase_ : str = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=A , cache_dir=model_args.cache_dir , ) else: logger.info('''Training new model from scratch''' ) UpperCAmelCase_ : int = AutoModelWithLMHead.from_config(A ) model.resize_token_embeddings(len(A ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( '''BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the''' '''--mlm flag (masked language modeling).''' ) if data_args.block_size <= 0: UpperCAmelCase_ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: UpperCAmelCase_ : Dict = min(data_args.block_size , tokenizer.max_len ) # Get datasets UpperCAmelCase_ : str = ( get_dataset(A , tokenizer=A , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) UpperCAmelCase_ : Any = ( get_dataset(A , tokenizer=A , evaluate=A , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": UpperCAmelCase_ : Optional[int] = DataCollatorForPermutationLanguageModeling( tokenizer=A , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: UpperCAmelCase_ : Tuple = DataCollatorForWholeWordMask( tokenizer=A , mlm_probability=data_args.mlm_probability ) else: UpperCAmelCase_ : List[str] = DataCollatorForLanguageModeling( tokenizer=A , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer UpperCAmelCase_ : Any = Trainer( model=A , args=A , data_collator=A , train_dataset=A , eval_dataset=A , prediction_loss_only=A , ) # Training if training_args.do_train: UpperCAmelCase_ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=A ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCAmelCase_ : Tuple = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) UpperCAmelCase_ : Dict = trainer.evaluate() UpperCAmelCase_ : Union[str, Any] = math.exp(eval_output['''eval_loss'''] ) UpperCAmelCase_ : Optional[int] = {'''perplexity''': perplexity} UpperCAmelCase_ : int = os.path.join(training_args.output_dir , '''eval_results_lm.txt''' ) if trainer.is_world_master(): with open(A , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key in sorted(result.keys() ): logger.info(''' %s = %s''' , A , str(result[key] ) ) writer.write('''%s = %s\n''' % (key, str(result[key] )) ) results.update(A ) return results def __UpperCAmelCase ( A : Tuple ) -> Tuple: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
304
1
def lowercase__ ( __snake_case : Tuple , __snake_case : List[Any] ): '''simple docstring''' UpperCAmelCase_ : Any = '' for i in table: res += inp[i - 1] return res def lowercase__ ( __snake_case : Optional[int] ): '''simple docstring''' return data[1:] + data[0] def lowercase__ ( __snake_case : Union[str, Any] , __snake_case : Optional[int] ): '''simple docstring''' UpperCAmelCase_ : Any = '' for i in range(len(__snake_case ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def lowercase__ ( __snake_case : Dict , __snake_case : Optional[int] ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = int('0b' + data[0] + data[-1] , 2 ) UpperCAmelCase_ : Any = int('0b' + data[1:3] , 2 ) return bin(s[row][col] )[2:] def lowercase__ ( __snake_case : Tuple , __snake_case : List[Any] , __snake_case : List[Any] , __snake_case : Optional[int] , __snake_case : str ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = message[:4] UpperCAmelCase_ : Optional[int] = message[4:] UpperCAmelCase_ : Union[str, Any] = apply_table(__snake_case , __snake_case ) UpperCAmelCase_ : Union[str, Any] = xor(__snake_case , __snake_case ) UpperCAmelCase_ : Dict = apply_sbox(__snake_case , temp[:4] ) # noqa: E741 UpperCAmelCase_ : Any = apply_sbox(__snake_case , temp[4:] ) UpperCAmelCase_ : str = '0' * (2 - len(__snake_case )) + l # noqa: E741 UpperCAmelCase_ : int = '0' * (2 - len(__snake_case )) + r UpperCAmelCase_ : int = apply_table(l + r , __snake_case ) UpperCAmelCase_ : int = xor(__snake_case , __snake_case ) return temp + right if __name__ == "__main__": __UpperCAmelCase = input('Enter 10 bit key: ') __UpperCAmelCase = input('Enter 8 bit message: ') __UpperCAmelCase = [6, 3, 7, 4, 8, 5, 10, 9] __UpperCAmelCase = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] __UpperCAmelCase = [2, 4, 3, 1] __UpperCAmelCase = [2, 6, 3, 1, 4, 8, 5, 7] __UpperCAmelCase = [4, 1, 3, 5, 7, 2, 8, 6] __UpperCAmelCase = [4, 1, 2, 3, 2, 3, 4, 1] __UpperCAmelCase = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] __UpperCAmelCase = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation __UpperCAmelCase = apply_table(key, paa_table) __UpperCAmelCase = temp[:5] __UpperCAmelCase = temp[5:] __UpperCAmelCase = left_shift(left) __UpperCAmelCase = left_shift(right) __UpperCAmelCase = apply_table(left + right, pa_table) __UpperCAmelCase = left_shift(left) __UpperCAmelCase = left_shift(right) __UpperCAmelCase = left_shift(left) __UpperCAmelCase = left_shift(right) __UpperCAmelCase = apply_table(left + right, pa_table) # encryption __UpperCAmelCase = apply_table(message, IP) __UpperCAmelCase = function(expansion, sa, sa, keya, temp) __UpperCAmelCase = temp[4:] + temp[:4] __UpperCAmelCase = function(expansion, sa, sa, keya, temp) __UpperCAmelCase = apply_table(temp, IP_inv) print('Cipher text is:', CT) # decryption __UpperCAmelCase = apply_table(CT, IP) __UpperCAmelCase = function(expansion, sa, sa, keya, temp) __UpperCAmelCase = temp[4:] + temp[:4] __UpperCAmelCase = function(expansion, sa, sa, keya, temp) __UpperCAmelCase = apply_table(temp, IP_inv) print('Plain text after decypting is:', PT)
145
def lowercase__ ( __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : Tuple = [1] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Tuple = 0, 0, 0 UpperCAmelCase_ : Union[str, Any] = ugly_nums[ia] * 2 UpperCAmelCase_ : Tuple = ugly_nums[ia] * 3 UpperCAmelCase_ : Union[str, Any] = ugly_nums[ia] * 5 for _ in range(1 , __snake_case ): UpperCAmelCase_ : Tuple = min(__snake_case , __snake_case , __snake_case ) ugly_nums.append(__snake_case ) if next_num == next_a: ia += 1 UpperCAmelCase_ : Union[str, Any] = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 UpperCAmelCase_ : Any = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 UpperCAmelCase_ : List[str] = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(F'{ugly_numbers(200) = }')
145
1
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class lowerCamelCase : """simple docstring""" @staticmethod def __A ( *__magic_name__ : Tuple , **__magic_name__ : List[Any] ) -> Optional[int]: pass def a__ ( __UpperCamelCase ): return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. A : int = ( "https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png" ) @is_pipeline_test @require_torch @require_vision class lowerCamelCase (unittest.TestCase ): """simple docstring""" lowerCamelCase__ = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def __A ( self : Optional[Any] , __magic_name__ : Optional[Any] , __magic_name__ : List[str] , __magic_name__ : Dict ) -> Tuple: SCREAMING_SNAKE_CASE_ = pipeline( "document-question-answering" , model=__magic_name__ , tokenizer=__magic_name__ , image_processor=__magic_name__ ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = list(zip(*apply_tesseract(load_image(__magic_name__ ) , __magic_name__ , "" ) ) ) SCREAMING_SNAKE_CASE_ = "What is the placebo?" SCREAMING_SNAKE_CASE_ = [ { "image": load_image(__magic_name__ ), "question": question, }, { "image": image, "question": question, }, { "image": image, "question": question, "word_boxes": word_boxes, }, ] return dqa_pipeline, examples def __A ( self : Union[str, Any] , __magic_name__ : Any , __magic_name__ : str ) -> Any: SCREAMING_SNAKE_CASE_ = dqa_pipeline(__magic_name__ , top_k=2 ) self.assertEqual( __magic_name__ , [ [ {"score": ANY(__magic_name__ ), "answer": ANY(__magic_name__ ), "start": ANY(__magic_name__ ), "end": ANY(__magic_name__ )}, {"score": ANY(__magic_name__ ), "answer": ANY(__magic_name__ ), "start": ANY(__magic_name__ ), "end": ANY(__magic_name__ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def __A ( self : List[Any] ) -> Tuple: SCREAMING_SNAKE_CASE_ = pipeline("document-question-answering" , model="hf-internal-testing/tiny-random-layoutlmv2" ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = "How many cats are there?" SCREAMING_SNAKE_CASE_ = [ {"score": 0.0001, "answer": "oy 2312/2019", "start": 38, "end": 39}, {"score": 0.0001, "answer": "oy 2312/2019 DUE", "start": 38, "end": 40}, ] SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual(nested_simplify(__magic_name__ , decimals=4 ) , __magic_name__ ) SCREAMING_SNAKE_CASE_ = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual(nested_simplify(__magic_name__ , decimals=4 ) , __magic_name__ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably SCREAMING_SNAKE_CASE_ = "./tests/fixtures/tests_samples/COCO/000000039769.png" SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual(__magic_name__ , [] ) # We can optionnally pass directly the words and bounding boxes SCREAMING_SNAKE_CASE_ = "./tests/fixtures/tests_samples/COCO/000000039769.png" SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , words=__magic_name__ , boxes=__magic_name__ , top_k=2 ) self.assertEqual(__magic_name__ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def __A ( self : str ) -> str: SCREAMING_SNAKE_CASE_ = pipeline( "document-question-answering" , model="tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa" , revision="9977165" , ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = "What is the invoice number?" SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.9944, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0009, "answer": "us-001", "start": 16, "end": 16}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.9944, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0009, "answer": "us-001", "start": 16, "end": 16}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ [ {"score": 0.9944, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0009, "answer": "us-001", "start": 16, "end": 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def __A ( self : Tuple ) -> int: SCREAMING_SNAKE_CASE_ = pipeline( "document-question-answering" , model="tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa" , revision="9977165" , max_seq_len=50 , ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = "What is the invoice number?" SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.9974, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.9948, "answer": "us-001", "start": 16, "end": 16}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.9974, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.9948, "answer": "us-001", "start": 16, "end": 16}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ [ {"score": 0.9974, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.9948, "answer": "us-001", "start": 16, "end": 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def __A ( self : List[str] ) -> Any: SCREAMING_SNAKE_CASE_ = AutoTokenizer.from_pretrained( "impira/layoutlm-document-qa" , revision="3dc6de3" , add_prefix_space=__magic_name__ ) SCREAMING_SNAKE_CASE_ = pipeline( "document-question-answering" , model="impira/layoutlm-document-qa" , tokenizer=__magic_name__ , revision="3dc6de3" , ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = "What is the invoice number?" SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.4251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0819, "answer": "1110212019", "start": 23, "end": 23}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.4251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0819, "answer": "1110212019", "start": 23, "end": 23}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ [ {"score": 0.4251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0819, "answer": "1110212019", "start": 23, "end": 23}, ] ] * 2 , ) SCREAMING_SNAKE_CASE_ = list(zip(*apply_tesseract(load_image(__magic_name__ ) , __magic_name__ , "" ) ) ) # This model should also work if `image` is set to None SCREAMING_SNAKE_CASE_ = dqa_pipeline({"image": None, "word_boxes": word_boxes, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.4251, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.0819, "answer": "1110212019", "start": 23, "end": 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def __A ( self : Union[str, Any] ) -> int: SCREAMING_SNAKE_CASE_ = AutoTokenizer.from_pretrained( "impira/layoutlm-document-qa" , revision="3dc6de3" , add_prefix_space=__magic_name__ ) SCREAMING_SNAKE_CASE_ = pipeline( "document-question-answering" , model="impira/layoutlm-document-qa" , tokenizer=__magic_name__ , revision="3dc6de3" , max_seq_len=50 , ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = "What is the invoice number?" SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.9999, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.9998, "answer": "us-001", "start": 16, "end": 16}, ] , ) SCREAMING_SNAKE_CASE_ = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ [ {"score": 0.9999, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.9998, "answer": "us-001", "start": 16, "end": 16}, ] ] * 2 , ) SCREAMING_SNAKE_CASE_ = list(zip(*apply_tesseract(load_image(__magic_name__ ) , __magic_name__ , "" ) ) ) # This model should also work if `image` is set to None SCREAMING_SNAKE_CASE_ = dqa_pipeline({"image": None, "word_boxes": word_boxes, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(__magic_name__ , decimals=4 ) , [ {"score": 0.9999, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.9998, "answer": "us-001", "start": 16, "end": 16}, ] , ) @slow @require_torch def __A ( self : Tuple ) -> List[str]: SCREAMING_SNAKE_CASE_ = pipeline( "document-question-answering" , model="naver-clova-ix/donut-base-finetuned-docvqa" , tokenizer=AutoTokenizer.from_pretrained("naver-clova-ix/donut-base-finetuned-docvqa" ) , feature_extractor="naver-clova-ix/donut-base-finetuned-docvqa" , ) SCREAMING_SNAKE_CASE_ = INVOICE_URL SCREAMING_SNAKE_CASE_ = "What is the invoice number?" SCREAMING_SNAKE_CASE_ = dqa_pipeline(image=__magic_name__ , question=__magic_name__ , top_k=2 ) self.assertEqual(nested_simplify(__magic_name__ , decimals=4 ) , [{"answer": "us-001"}] ) @require_tf @unittest.skip("Document question answering not implemented in TF" ) def __A ( self : Optional[Any] ) -> Union[str, Any]: pass
118
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCamelCase : """simple docstring""" def __init__( self : List[Any] , __magic_name__ : Any , __magic_name__ : List[Any]=13 , __magic_name__ : List[Any]=2 , __magic_name__ : Tuple=24 , __magic_name__ : List[str]=16 , __magic_name__ : Dict=True , __magic_name__ : List[Any]=True , __magic_name__ : Optional[int]=32 , __magic_name__ : Tuple=5 , __magic_name__ : int=4 , __magic_name__ : Tuple=37 , __magic_name__ : List[str]="gelu" , __magic_name__ : Tuple=0.1 , __magic_name__ : Tuple=0.1 , __magic_name__ : Union[str, Any]=10 , __magic_name__ : Tuple=0.02 , __magic_name__ : Tuple=None , __magic_name__ : Any=2 , __magic_name__ : Dict=2 , ) -> int: SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = patch_size SCREAMING_SNAKE_CASE_ = max_length SCREAMING_SNAKE_CASE_ = num_mel_bins SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = use_labels SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = type_sequence_label_size SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = scope SCREAMING_SNAKE_CASE_ = frequency_stride SCREAMING_SNAKE_CASE_ = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) SCREAMING_SNAKE_CASE_ = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 SCREAMING_SNAKE_CASE_ = (self.max_length - self.patch_size) // self.time_stride + 1 SCREAMING_SNAKE_CASE_ = frequency_out_dimension * time_out_dimension SCREAMING_SNAKE_CASE_ = num_patches + 2 def __A ( self : Any ) -> Any: SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) SCREAMING_SNAKE_CASE_ = None if self.use_labels: SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ = self.get_config() return config, input_values, labels def __A ( self : Any ) -> Dict: return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__magic_name__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def __A ( self : List[Any] , __magic_name__ : List[str] , __magic_name__ : str , __magic_name__ : List[str] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = ASTModel(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() SCREAMING_SNAKE_CASE_ = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self : Tuple ) -> str: SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ) = config_and_inputs SCREAMING_SNAKE_CASE_ = {"input_values": input_values} return config, inputs_dict @require_torch class lowerCamelCase (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowerCamelCase__ = ( {'''audio-classification''': ASTForAudioClassification, '''feature-extraction''': ASTModel} if is_torch_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def __A ( self : Tuple , __magic_name__ : Union[str, Any] , __magic_name__ : Optional[Any] , __magic_name__ : Dict , __magic_name__ : Optional[int] , __magic_name__ : Tuple ) -> Tuple: if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def __A ( self : Optional[Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = ASTModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=__magic_name__ , has_text_modality=__magic_name__ , hidden_size=37 ) def __A ( self : Union[str, Any] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="AST does not use inputs_embeds" ) def __A ( self : Optional[Any] ) -> Tuple: pass def __A ( self : int ) -> int: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(__magic_name__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__magic_name__ , nn.Linear ) ) def __A ( self : List[Any] ) -> str: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(__magic_name__ ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = ["input_values"] self.assertListEqual(arg_names[:1] , __magic_name__ ) def __A ( self : int ) -> Any: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) @slow def __A ( self : int ) -> int: for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_ = ASTModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) def a__ ( ): SCREAMING_SNAKE_CASE_ = hf_hub_download( repo_id="nielsr/audio-spectogram-transformer-checkpoint" , filename="sample_audio.flac" , repo_type="dataset" ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = torchaudio.load(__UpperCamelCase ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCamelCase (unittest.TestCase ): """simple docstring""" @cached_property def __A ( self : List[Any] ) -> List[Any]: return ( ASTFeatureExtractor.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ) if is_torchaudio_available() else None ) @slow def __A ( self : Union[str, Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE_ = self.default_feature_extractor SCREAMING_SNAKE_CASE_ = ASTForAudioClassification.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ).to(__magic_name__ ) SCREAMING_SNAKE_CASE_ = self.default_feature_extractor SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = prepare_audio() SCREAMING_SNAKE_CASE_ = audio.squeeze().numpy() SCREAMING_SNAKE_CASE_ = feature_extractor(__magic_name__ , sampling_rate=__magic_name__ , return_tensors="pt" ).to(__magic_name__ ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**__magic_name__ ) # verify the logits SCREAMING_SNAKE_CASE_ = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , __magic_name__ ) SCREAMING_SNAKE_CASE_ = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(__magic_name__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __magic_name__ , atol=1e-4 ) )
118
1
'''simple docstring''' def UpperCamelCase ( _lowerCamelCase : list[list[int]] , _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : list[int] ): # 1. Validate that path exists between current and next vertices if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def UpperCamelCase ( _lowerCamelCase : list[list[int]] , _lowerCamelCase : list[int] , _lowerCamelCase : int ): # Base Case if curr_ind == len(_lowerCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(_lowerCamelCase ) ): if valid_connection(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): # Insert current vertex into path as next transition A__ = next_ver # Validate created path if util_hamilton_cycle(_lowerCamelCase , _lowerCamelCase , curr_ind + 1 ): return True # Backtrack A__ = -1 return False def UpperCamelCase ( _lowerCamelCase : list[list[int]] , _lowerCamelCase : int = 0 ): A__ = [-1] * (len(_lowerCamelCase ) + 1) # initialize start and end of path with starting index A__ = A__ = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(_lowerCamelCase , _lowerCamelCase , 1 ) else []
123
'''simple docstring''' import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCAmelCase ( UpperCamelCase__ ): __lowercase = (PNDMScheduler,) __lowercase = (("""num_inference_steps""", 50),) def UpperCAmelCase_ ( self :List[str] , **lowercase_ :Optional[int] )-> List[str]: A__ = { "num_train_timesteps": 10_00, "beta_start": 0.0_0_0_1, "beta_end": 0.0_2, "beta_schedule": "linear", } config.update(**lowercase_ ) return config def UpperCAmelCase_ ( self :int , lowercase_ :Optional[int]=0 , **lowercase_ :Any )-> int: A__ = dict(self.forward_default_kwargs ) A__ = kwargs.pop("num_inference_steps" , lowercase_ ) A__ = self.dummy_sample A__ = 0.1 * sample A__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: A__ = self.get_scheduler_config(**lowercase_ ) A__ = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals A__ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) A__ = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals A__ = dummy_past_residuals[:] A__ = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample A__ = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" A__ = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample A__ = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCAmelCase_ ( self :List[Any] )-> Tuple: pass def UpperCAmelCase_ ( self :Optional[Any] , lowercase_ :Optional[Any]=0 , **lowercase_ :List[str] )-> str: A__ = dict(self.forward_default_kwargs ) A__ = kwargs.pop("num_inference_steps" , lowercase_ ) A__ = self.dummy_sample A__ = 0.1 * sample A__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: A__ = self.get_scheduler_config() A__ = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) A__ = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) A__ = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) A__ = dummy_past_residuals[:] A__ = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample A__ = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" A__ = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample A__ = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCAmelCase_ ( self :Optional[int] , **lowercase_ :Dict )-> Optional[int]: A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(**lowercase_ ) A__ = scheduler_class(**lowercase_ ) A__ = 10 A__ = self.dummy_model() A__ = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): A__ = model(lowercase_ , lowercase_ ) A__ = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): A__ = model(lowercase_ , lowercase_ ) A__ = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ ).prev_sample return sample def UpperCAmelCase_ ( self :int )-> Union[str, Any]: A__ = dict(self.forward_default_kwargs ) A__ = kwargs.pop("num_inference_steps" , lowercase_ ) for scheduler_class in self.scheduler_classes: A__ = self.get_scheduler_config() A__ = scheduler_class(**lowercase_ ) A__ = self.dummy_sample A__ = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase_ , "set_timesteps" ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_ , "set_timesteps" ): A__ = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) A__ = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] A__ = dummy_past_residuals[:] A__ = scheduler.step_prk(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample A__ = scheduler.step_prk(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) A__ = scheduler.step_plms(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample A__ = scheduler.step_plms(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCAmelCase_ ( self :Any )-> Dict: for timesteps in [1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowercase_ ) def UpperCAmelCase_ ( self :List[str] )-> int: for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(steps_offset=1 ) A__ = scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [9_01, 8_51, 8_51, 8_01, 8_01, 7_51, 7_51, 7_01, 7_01, 6_51, 6_51, 6_01, 6_01, 5_01, 4_01, 3_01, 2_01, 1_01, 1] ) , ) def UpperCAmelCase_ ( self :Optional[Any] )-> Any: for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1] , [0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=lowercase_ , beta_end=lowercase_ ) def UpperCAmelCase_ ( self :Tuple )-> Tuple: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def UpperCAmelCase_ ( self :Optional[Any] )-> Tuple: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def UpperCAmelCase_ ( self :Optional[int] )-> Tuple: for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def UpperCAmelCase_ ( self :List[Any] )-> Optional[int]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 1_00] ): self.check_over_forward(num_inference_steps=lowercase_ ) def UpperCAmelCase_ ( self :List[str] )-> str: # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 A__ = 27 for scheduler_class in self.scheduler_classes: A__ = self.dummy_sample A__ = 0.1 * sample A__ = self.get_scheduler_config() A__ = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): A__ = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample def UpperCAmelCase_ ( self :List[str] )-> str: with self.assertRaises(lowercase_ ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def UpperCAmelCase_ ( self :int )-> str: A__ = self.full_loop() A__ = torch.sum(torch.abs(lowercase_ ) ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_9_8.1_3_1_8 ) < 1E-2 assert abs(result_mean.item() - 0.2_5_8_0 ) < 1E-3 def UpperCAmelCase_ ( self :Optional[Any] )-> List[Any]: A__ = self.full_loop(prediction_type="v_prediction" ) A__ = torch.sum(torch.abs(lowercase_ ) ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 6_7.3_9_8_6 ) < 1E-2 assert abs(result_mean.item() - 0.0_8_7_8 ) < 1E-3 def UpperCAmelCase_ ( self :Tuple )-> Tuple: # We specify different beta, so that the first alpha is 0.99 A__ = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.0_1 ) A__ = torch.sum(torch.abs(lowercase_ ) ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 2_3_0.0_3_9_9 ) < 1E-2 assert abs(result_mean.item() - 0.2_9_9_5 ) < 1E-3 def UpperCAmelCase_ ( self :Dict )-> Any: # We specify different beta, so that the first alpha is 0.99 A__ = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.0_1 ) A__ = torch.sum(torch.abs(lowercase_ ) ) A__ = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_8_6.9_4_8_2 ) < 1E-2 assert abs(result_mean.item() - 0.2_4_3_4 ) < 1E-3
123
1
"""simple docstring""" import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def _UpperCAmelCase ( self ) -> int: debug_launcher(test_script.main ) def _UpperCAmelCase ( self ) -> List[str]: debug_launcher(test_ops.main )
320
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline __A = "path-to-your-trained-model" __A = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("cuda") __A = "A photo of sks dog in a bucket" __A = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save("dog-bucket.png")
148
0
def _SCREAMING_SNAKE_CASE ( lowercase : str , lowercase : int ): '''simple docstring''' return [sentence[i : i + ngram_size] for i in range(len(lowercase ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
208
import os import time import numpy as np import onnxruntime as ort lowerCamelCase : int = "1" lowerCamelCase : int = "0" lowerCamelCase : Union[str, Any] = "1" lowerCamelCase : List[Any] = ort.SessionOptions() lowerCamelCase : Optional[Any] = ort.GraphOptimizationLevel.ORT_DISABLE_ALL print("Create inference session...") lowerCamelCase : Union[str, Any] = ["TensorrtExecutionProvider", "CUDAExecutionProvider"] lowerCamelCase : Tuple = ort.InferenceSession("model.onnx", sess_options=sess_opt, providers=execution_provider) lowerCamelCase : List[Any] = ort.RunOptions() lowerCamelCase : List[str] = 128 lowerCamelCase : List[Any] = 1 lowerCamelCase : Union[str, Any] = np.ones((batch, sequence), dtype=np.intaa) lowerCamelCase : Dict = np.ones((batch, sequence), dtype=np.intaa) lowerCamelCase : Optional[Any] = np.ones((batch, sequence), dtype=np.intaa) print("Warm up phase...") sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print("Start inference...") lowerCamelCase : int = time.time() lowerCamelCase : Dict = 2_000 lowerCamelCase : Any = {} for iter in range(max_iters): lowerCamelCase : Union[str, Any] = sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print("Average Inference Time = {:.3f} ms".format((time.time() - start_time) * 1_000 / max_iters))
208
1
import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowerCAmelCase_ ( _lowercase : Optional[Any] , _lowercase : Optional[Any]) -> Any: """simple docstring""" assert isinstance(_lowercase , _lowercase) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("""keep_in_memory""" , [False, True]) def lowerCAmelCase_ ( _lowercase : Dict , _lowercase : Any , _lowercase : Any) -> Any: """simple docstring""" a__ : Any = tmp_path / """cache""" a__ : Dict = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): a__ : Union[str, Any] = ParquetDatasetReader(_lowercase , cache_dir=_lowercase , keep_in_memory=_lowercase).read() _check_parquet_dataset(_lowercase , _lowercase) @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCAmelCase_ ( _lowercase : Union[str, Any] , _lowercase : List[Any] , _lowercase : Dict) -> Union[str, Any]: """simple docstring""" a__ : Optional[int] = tmp_path / """cache""" a__ : Dict = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} a__ : Tuple = features.copy() if features else default_expected_features a__ : Optional[int] = ( Features({feature: Value(_lowercase) for feature, dtype in features.items()}) if features is not None else None ) a__ : Any = ParquetDatasetReader(_lowercase , features=_lowercase , cache_dir=_lowercase).read() _check_parquet_dataset(_lowercase , _lowercase) @pytest.mark.parametrize("""split""" , [None, NamedSplit("""train"""), """train""", """test"""]) def lowerCAmelCase_ ( _lowercase : Optional[int] , _lowercase : Optional[Any] , _lowercase : Union[str, Any]) -> Optional[int]: """simple docstring""" a__ : int = tmp_path / """cache""" a__ : Tuple = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} a__ : Tuple = ParquetDatasetReader(_lowercase , cache_dir=_lowercase , split=_lowercase).read() _check_parquet_dataset(_lowercase , _lowercase) assert dataset.split == split if split else "train" @pytest.mark.parametrize("""path_type""" , [str, list]) def lowerCAmelCase_ ( _lowercase : List[str] , _lowercase : Dict , _lowercase : Optional[Any]) -> Union[str, Any]: """simple docstring""" if issubclass(_lowercase , _lowercase): a__ : List[Any] = parquet_path elif issubclass(_lowercase , _lowercase): a__ : Optional[Any] = [parquet_path] a__ : List[Any] = tmp_path / """cache""" a__ : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} a__ : Optional[Any] = ParquetDatasetReader(_lowercase , cache_dir=_lowercase).read() _check_parquet_dataset(_lowercase , _lowercase) def lowerCAmelCase_ ( _lowercase : Optional[Any] , _lowercase : Optional[Any] , _lowercase : Optional[Any]=("train",)) -> List[str]: """simple docstring""" assert isinstance(_lowercase , _lowercase) for split in splits: a__ : Union[str, Any] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("""keep_in_memory""" , [False, True]) def lowerCAmelCase_ ( _lowercase : str , _lowercase : Any , _lowercase : Dict) -> Dict: """simple docstring""" a__ : str = tmp_path / """cache""" a__ : Tuple = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): a__ : Optional[int] = ParquetDatasetReader( {"""train""": parquet_path} , cache_dir=_lowercase , keep_in_memory=_lowercase).read() _check_parquet_datasetdict(_lowercase , _lowercase) @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def lowerCAmelCase_ ( _lowercase : Tuple , _lowercase : List[str] , _lowercase : List[Any]) -> List[str]: """simple docstring""" a__ : List[str] = tmp_path / """cache""" a__ : List[str] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} a__ : Tuple = features.copy() if features else default_expected_features a__ : Optional[int] = ( Features({feature: Value(_lowercase) for feature, dtype in features.items()}) if features is not None else None ) a__ : Dict = ParquetDatasetReader({"""train""": parquet_path} , features=_lowercase , cache_dir=_lowercase).read() _check_parquet_datasetdict(_lowercase , _lowercase) @pytest.mark.parametrize("""split""" , [None, NamedSplit("""train"""), """train""", """test"""]) def lowerCAmelCase_ ( _lowercase : List[Any] , _lowercase : Optional[Any] , _lowercase : Union[str, Any]) -> str: """simple docstring""" if split: a__ : str = {split: parquet_path} else: a__ : Union[str, Any] = """train""" a__ : Dict = {"""train""": parquet_path, """test""": parquet_path} a__ : Optional[int] = tmp_path / """cache""" a__ : int = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} a__ : Dict = ParquetDatasetReader(_lowercase , cache_dir=_lowercase).read() _check_parquet_datasetdict(_lowercase , _lowercase , splits=list(path.keys())) assert all(dataset[split].split == split for split in path.keys()) def lowerCAmelCase_ ( _lowercase : Any , _lowercase : Union[str, Any]) -> List[str]: """simple docstring""" a__ : Tuple = ParquetDatasetWriter(_lowercase , tmp_path / """foo.parquet""") assert writer.write() > 0 a__ : Tuple = pq.ParquetFile(tmp_path / """foo.parquet""") a__ : List[Any] = pf.read() assert dataset.data.table == output_table def lowerCAmelCase_ ( _lowercase : Union[str, Any] , _lowercase : str) -> Dict: """simple docstring""" a__ : List[Any] = str(shared_datadir / """test_image_rgb.jpg""") a__ : Union[str, Any] = {"""image""": [image_path]} a__ : List[Any] = Features({"""image""": Image()}) a__ : Union[str, Any] = Dataset.from_dict(_lowercase , features=_lowercase) a__ : Optional[int] = ParquetDatasetWriter(_lowercase , tmp_path / """foo.parquet""") assert writer.write() > 0 a__ : List[str] = Dataset.from_parquet(str(tmp_path / """foo.parquet""")) assert dataset.features == reloaded_dataset.features a__ : str = ParquetDatasetReader(str(tmp_path / """foo.parquet""") , streaming=_lowercase).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( """feature, expected""" , [ (Features({"""foo""": Value("""int32""")}), None), (Features({"""image""": Image(), """foo""": Value("""int32""")}), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({"""nested""": Sequence(Audio())}), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ] , ) def lowerCAmelCase_ ( _lowercase : Optional[Any] , _lowercase : List[Any]) -> str: """simple docstring""" assert get_writer_batch_size(_lowercase) == expected
170
from __future__ import annotations import math def lowerCAmelCase_ ( _lowercase : int) -> bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_lowercase) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True _lowercase : str =[num for num in range(3, 10_0001, 2) if not is_prime(num)] def lowerCAmelCase_ ( _lowercase : int) -> list[int]: """simple docstring""" if not isinstance(_lowercase , _lowercase): raise ValueError("""n must be an integer""") if n <= 0: raise ValueError("""n must be >= 0""") a__ : int = [] for num in range(len(_lowercase)): a__ : Any = 0 while 2 * i * i <= odd_composites[num]: a__ : Union[str, Any] = odd_composites[num] - 2 * i * i if is_prime(_lowercase): break i += 1 else: list_nums.append(odd_composites[num]) if len(_lowercase) == n: return list_nums return [] def lowerCAmelCase_ ( ) -> int: """simple docstring""" return compute_nums(1)[0] if __name__ == "__main__": print(f'{solution() = }')
170
1
'''simple docstring''' import math def __magic_name__ ( A , A ) -> List[str]: if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(A ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError('This should never happen' ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. lowerCAmelCase_ = "Enter the base and the power separated by a comma: " lowerCAmelCase_ , lowerCAmelCase_ = map(int, input(prompt).split(",")) lowerCAmelCase_ , lowerCAmelCase_ = map(int, input(prompt).split(",")) # We find the log of each number, using the function res(), which takes two # arguments. lowerCAmelCase_ = res(xa, ya) lowerCAmelCase_ = res(xa, ya) # We check for the largest number if resa > resa: print("Largest number is", xa, "^", ya) elif resa > resa: print("Largest number is", xa, "^", ya) else: print("Both are equal")
332
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class lowerCamelCase ( __lowerCAmelCase ): snake_case_ = 42 class lowerCamelCase ( __lowerCAmelCase , __lowerCAmelCase ): @register_to_config def __init__( self, lowercase_ = 3, lowercase_ = 3, lowercase_ = ("DownEncoderBlock2D",), lowercase_ = ("UpDecoderBlock2D",), lowercase_ = (64,), lowercase_ = 1, lowercase_ = "silu", lowercase_ = 3, lowercase_ = 32, lowercase_ = 256, lowercase_ = 32, lowercase_ = None, lowercase_ = 0.18_215, lowercase_ = "group", ) -> str: super().__init__() # pass init params to Encoder snake_case = Encoder( in_channels=lowercase_, out_channels=lowercase_, down_block_types=lowercase_, block_out_channels=lowercase_, layers_per_block=lowercase_, act_fn=lowercase_, norm_num_groups=lowercase_, double_z=lowercase_, ) snake_case = vq_embed_dim if vq_embed_dim is not None else latent_channels snake_case = nn.Convad(lowercase_, lowercase_, 1 ) snake_case = VectorQuantizer(lowercase_, lowercase_, beta=0.25, remap=lowercase_, sane_index_shape=lowercase_ ) snake_case = nn.Convad(lowercase_, lowercase_, 1 ) # pass init params to Decoder snake_case = Decoder( in_channels=lowercase_, out_channels=lowercase_, up_block_types=lowercase_, block_out_channels=lowercase_, layers_per_block=lowercase_, act_fn=lowercase_, norm_num_groups=lowercase_, norm_type=lowercase_, ) @apply_forward_hook def _lowerCamelCase ( self, lowercase_, lowercase_ = True ) -> VQEncoderOutput: snake_case = self.encoder(lowercase_ ) snake_case = self.quant_conv(lowercase_ ) if not return_dict: return (h,) return VQEncoderOutput(latents=lowercase_ ) @apply_forward_hook def _lowerCamelCase ( self, lowercase_, lowercase_ = False, lowercase_ = True ) -> Union[DecoderOutput, torch.FloatTensor]: # also go through quantization layer if not force_not_quantize: snake_case , snake_case , snake_case = self.quantize(lowercase_ ) else: snake_case = h snake_case = self.post_quant_conv(lowercase_ ) snake_case = self.decoder(lowercase_, quant if self.config.norm_type == 'spatial' else None ) if not return_dict: return (dec,) return DecoderOutput(sample=lowercase_ ) def _lowerCamelCase ( self, lowercase_, lowercase_ = True ) -> Union[DecoderOutput, torch.FloatTensor]: snake_case = sample snake_case = self.encode(lowercase_ ).latents snake_case = self.decode(lowercase_ ).sample if not return_dict: return (dec,) return DecoderOutput(sample=lowercase_ )
332
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCAmelCase = { '''configuration_transfo_xl''': ['''TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TransfoXLConfig'''], '''tokenization_transfo_xl''': ['''TransfoXLCorpus''', '''TransfoXLTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ '''TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AdaptiveEmbedding''', '''TransfoXLForSequenceClassification''', '''TransfoXLLMHeadModel''', '''TransfoXLModel''', '''TransfoXLPreTrainedModel''', '''load_tf_weights_in_transfo_xl''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ '''TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFAdaptiveEmbedding''', '''TFTransfoXLForSequenceClassification''', '''TFTransfoXLLMHeadModel''', '''TFTransfoXLMainLayer''', '''TFTransfoXLModel''', '''TFTransfoXLPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
119
from __future__ import annotations from typing import Any class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_ ) -> None: UpperCamelCase : Any = num_of_nodes UpperCamelCase : list[list[int]] = [] UpperCamelCase : dict[int, int] = {} def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> None: self.m_edges.append([u_node, v_node, weight] ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> int: if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> None: if self.m_component[u_node] != u_node: for k in self.m_component: UpperCamelCase : Dict = self.find_component(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> None: if component_size[u_node] <= component_size[v_node]: UpperCamelCase : Tuple = v_node component_size[v_node] += component_size[u_node] self.set_component(SCREAMING_SNAKE_CASE_ ) elif component_size[u_node] >= component_size[v_node]: UpperCamelCase : Union[str, Any] = self.find_component(SCREAMING_SNAKE_CASE_ ) component_size[u_node] += component_size[v_node] self.set_component(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> None: UpperCamelCase : int = [] UpperCamelCase : int = 0 UpperCamelCase : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) UpperCamelCase : List[str] = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: UpperCamelCase , UpperCamelCase , UpperCamelCase : List[Any] = edge UpperCamelCase : str = self.m_component[u] UpperCamelCase : Any = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): UpperCamelCase : Any = [u, v, w] for edge in minimum_weight_edge: if isinstance(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): UpperCamelCase , UpperCamelCase , UpperCamelCase : int = edge UpperCamelCase : List[Any] = self.m_component[u] UpperCamelCase : Tuple = self.m_component[v] if u_component != v_component: mst_weight += w self.union(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) print(F"""Added edge [{u} - {v}]\nAdded weight: {w}\n""" ) num_of_components -= 1 UpperCamelCase : Optional[Any] = [-1] * self.m_num_of_nodes print(F"""The total weight of the minimal spanning tree is: {mst_weight}""" ) def UpperCamelCase ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
119
1
"""simple docstring""" from __future__ import annotations from collections.abc import Generator def lowerCamelCase () -> List[Any]: lowercase :Optional[Any] = {} lowercase :List[Any] = 2 while True: lowercase :List[Any] = factor_map.pop(a_ , a_) if factor: lowercase :List[Any] = factor + prime while x in factor_map: x += factor lowercase :Any = factor else: lowercase :Union[str, Any] = prime yield prime prime += 1 def lowerCamelCase (a_ :float = 1E10) -> Dict: lowercase :str = sieve() lowercase :str = 1 while True: lowercase :Any = next(a_) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(a_) n += 2 if __name__ == "__main__": print(solution())
353
"""simple docstring""" import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = '''▁''' UpperCAmelCase = {'''vocab_file''': '''vocab.txt''', '''sentencepiece_model_ckpt''': '''sentencepiece.bpe.model'''} UpperCAmelCase = { '''sentencepiece_model_file''': '''sentencepiece.bpe.model''', '''vocab_file''': '''vocab.txt''', } UpperCAmelCase = { '''vocab_file''': { '''ernie-m-base''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt''', '''ernie-m-large''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt''', }, '''sentencepiece_model_file''': { '''ernie-m-base''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model''', '''ernie-m-large''': '''https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model''', }, } UpperCAmelCase = { '''ernie-m-base''': 514, '''ernie-m-large''': 514, } UpperCAmelCase = { '''ernie-m-base''': {'''do_lower_case''': False}, '''ernie-m-large''': {'''do_lower_case''': False}, } class __magic_name__ ( __UpperCAmelCase ): __A : List[str] = ["input_ids"] __A : Optional[Any] = VOCAB_FILES_NAMES __A : str = PRETRAINED_INIT_CONFIGURATION __A : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A : List[str] = PRETRAINED_VOCAB_FILES_MAP __A : List[str] = RESOURCE_FILES_NAMES def __init__( self : Dict , snake_case__ : List[Any] , snake_case__ : List[Any]=None , snake_case__ : int=False , snake_case__ : Optional[int]="utf8" , snake_case__ : List[str]="[UNK]" , snake_case__ : Tuple="[SEP]" , snake_case__ : List[Any]="[PAD]" , snake_case__ : Dict="[CLS]" , snake_case__ : Dict="[MASK]" , snake_case__ : Optional[Dict[str, Any]] = None , **snake_case__ : str , ): '''simple docstring''' lowercase :Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , pad_token=snake_case__ , cls_token=snake_case__ , mask_token=snake_case__ , vocab_file=snake_case__ , encoding=snake_case__ , sp_model_kwargs=self.sp_model_kwargs , **snake_case__ , ) lowercase :Dict = do_lower_case lowercase :str = sentencepiece_model_ckpt lowercase :Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowercase :Tuple = self.load_vocab(filepath=snake_case__ ) else: lowercase :str = {self.sp_model.id_to_piece(snake_case__ ): id for id in range(self.sp_model.get_piece_size() )} lowercase :Any = {v: k for k, v in self.vocab.items()} def __snake_case ( self : List[str] , snake_case__ : str ): '''simple docstring''' if text is None: return None lowercase :List[Any] = self.tokenize(snake_case__ ) lowercase , lowercase :List[str] = '''''', [] for i, ch in enumerate(snake_case__ ): if ch in self.SP_CHAR_MAPPING: lowercase :Optional[int] = self.SP_CHAR_MAPPING.get(snake_case__ ) else: lowercase :Optional[int] = unicodedata.normalize('''NFKC''' , snake_case__ ) if self.is_whitespace(snake_case__ ): continue normalized_text += ch char_mapping.extend([i] * len(snake_case__ ) ) lowercase , lowercase , lowercase :int = normalized_text, [], 0 if self.do_lower_case: lowercase :Any = text.lower() for token in split_tokens: if token[:1] == "▁": lowercase :Tuple = token[1:] lowercase :List[str] = text[offset:].index(snake_case__ ) + offset lowercase :Tuple = start + len(snake_case__ ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowercase :int = end return token_mapping @property def __snake_case ( self : List[Any] ): '''simple docstring''' return len(self.vocab ) def __snake_case ( self : Optional[Any] ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self : Optional[int] ): '''simple docstring''' lowercase :Any = self.__dict__.copy() lowercase :Optional[int] = None return state def __setstate__( self : Tuple , snake_case__ : Dict ): '''simple docstring''' lowercase :Optional[int] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase :Dict = {} lowercase :List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def __snake_case ( self : int , snake_case__ : List[Any] ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(snake_case__ , snake_case__ ) for c in text) ) def __snake_case ( self : List[str] , snake_case__ : Optional[int] , snake_case__ : int=False , snake_case__ : Dict=6_4 , snake_case__ : Any=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get('''enable_sampling''' ) is True: lowercase :Any = True if self.sp_model_kwargs.get('''alpha''' ) is not None: lowercase :Any = self.sp_model_kwargs.get('''alpha''' ) if self.sp_model_kwargs.get('''nbest_size''' ) is not None: lowercase :Optional[Any] = self.sp_model_kwargs.get('''nbest_size''' ) if not enable_sampling: lowercase :Any = self.sp_model.EncodeAsPieces(snake_case__ ) else: lowercase :List[Any] = self.sp_model.SampleEncodeAsPieces(snake_case__ , snake_case__ , snake_case__ ) lowercase :str = [] for pi, piece in enumerate(snake_case__ ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(snake_case__ ) and pi != 0: new_pieces.append(snake_case__ ) continue else: continue lowercase :int = 0 for i, chunk in enumerate(snake_case__ ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(snake_case__ ) or self.is_punct(snake_case__ ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(snake_case__ ) lowercase :Optional[int] = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase :str = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowercase :Dict = i if len(snake_case__ ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def __snake_case ( self : Dict , snake_case__ : str ): '''simple docstring''' lowercase :int = ''''''.join(snake_case__ ).replace(snake_case__ , ''' ''' ).strip() return out_string def __snake_case ( self : int , snake_case__ : str ): '''simple docstring''' lowercase :Tuple = self.convert_ids_to_tokens(snake_case__ ) lowercase :Any = ''''''.join(snake_case__ ).replace(snake_case__ , ''' ''' ).strip() return out_string def __snake_case ( self : int , snake_case__ : Union[str, Any] ): '''simple docstring''' return self.vocab.get(snake_case__ , self.vocab.get(self.unk_token ) ) def __snake_case ( self : List[Any] , snake_case__ : List[str] ): '''simple docstring''' return self.reverse_vocab.get(snake_case__ , self.unk_token ) def __snake_case ( self : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : Any=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase :int = [self.cls_token_id] lowercase :str = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def __snake_case ( self : Any , snake_case__ : Dict , snake_case__ : str=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def __snake_case ( self : List[Any] , snake_case__ : Optional[int] , snake_case__ : Any=None , snake_case__ : Optional[int]=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(snake_case__ )) + [1, 1] + ([0] * len(snake_case__ )) + [1] return [1] + ([0] * len(snake_case__ )) + [1] def __snake_case ( self : List[Any] , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(snake_case__ ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(snake_case__ ) + 1) + [1] * (len(snake_case__ ) + 3) def __snake_case ( self : List[Any] , snake_case__ : Any ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def __snake_case ( self : List[str] , snake_case__ : Any ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def __snake_case ( self : List[str] , snake_case__ : Union[str, Any] ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def __snake_case ( self : Optional[int] , snake_case__ : List[str] ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(snake_case__ ) == 1: lowercase :str = unicodedata.category(snake_case__ ) if cat == "Zs": return True return False def __snake_case ( self : str , snake_case__ : Any ): '''simple docstring''' lowercase :Dict = {} with io.open(snake_case__ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(snake_case__ ): lowercase :Dict = line.rstrip('''\n''' ) lowercase :str = int(snake_case__ ) return token_to_idx def __snake_case ( self : Dict , snake_case__ : str , snake_case__ : Optional[str] = None ): '''simple docstring''' lowercase :Optional[int] = 0 if os.path.isdir(snake_case__ ): lowercase :str = os.path.join( snake_case__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) else: lowercase :Any = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory with open(snake_case__ , '''w''' , encoding='''utf-8''' ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda snake_case__ : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ''' Please check that the vocabulary is not corrupted!''' ) lowercase :Optional[int] = token_index writer.write(token + '''\n''' ) index += 1 lowercase :int = os.path.join(snake_case__ , '''sentencepiece.bpe.model''' ) with open(snake_case__ , '''wb''' ) as fi: lowercase :Tuple = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (vocab_file,)
172
0
from typing import Dict, Iterable, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract _SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) def UpperCAmelCase_ ( _A , _A , _A ): '''simple docstring''' return [ int(10_00 * (box[0] / width) ), int(10_00 * (box[1] / height) ), int(10_00 * (box[2] / width) ), int(10_00 * (box[3] / height) ), ] def UpperCAmelCase_ ( _A , _A , _A ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = to_pil_image(_A ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = pil_image.size SCREAMING_SNAKE_CASE__ = pytesseract.image_to_data(_A , lang=_A , output_type='''dict''' , config=_A ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates SCREAMING_SNAKE_CASE__ = [idx for idx, word in enumerate(_A ) if not word.strip()] SCREAMING_SNAKE_CASE__ = [word for idx, word in enumerate(_A ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE__ = [coord for idx, coord in enumerate(_A ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE__ = [coord for idx, coord in enumerate(_A ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE__ = [coord for idx, coord in enumerate(_A ) if idx not in irrelevant_indices] SCREAMING_SNAKE_CASE__ = [coord for idx, coord in enumerate(_A ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format SCREAMING_SNAKE_CASE__ = [] for x, y, w, h in zip(_A , _A , _A , _A ): SCREAMING_SNAKE_CASE__ = [x, y, x + w, y + h] actual_boxes.append(_A ) # finally, normalize the bounding boxes SCREAMING_SNAKE_CASE__ = [] for box in actual_boxes: normalized_boxes.append(normalize_box(_A , _A , _A ) ) assert len(_A ) == len(_A ), "Not as many words as there are bounding boxes" return words, normalized_boxes class UpperCAmelCase__ ( A__ ): """simple docstring""" a = ["pixel_values"] def __init__( self : Tuple , __lowerCamelCase : bool = True , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : PILImageResampling = PILImageResampling.BILINEAR , __lowerCamelCase : bool = True , __lowerCamelCase : float = 1 / 255 , __lowerCamelCase : bool = True , __lowerCamelCase : Union[float, Iterable[float]] = None , __lowerCamelCase : Union[float, Iterable[float]] = None , __lowerCamelCase : bool = True , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[str] = "" , **__lowerCamelCase : List[Any] , ) -> None: super().__init__(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = size if size is not None else {'''height''': 224, '''width''': 224} SCREAMING_SNAKE_CASE__ = get_size_dict(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = do_resize SCREAMING_SNAKE_CASE__ = size SCREAMING_SNAKE_CASE__ = resample SCREAMING_SNAKE_CASE__ = do_rescale SCREAMING_SNAKE_CASE__ = rescale_value SCREAMING_SNAKE_CASE__ = do_normalize SCREAMING_SNAKE_CASE__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE__ = image_std if image_std is not None else IMAGENET_STANDARD_STD SCREAMING_SNAKE_CASE__ = apply_ocr SCREAMING_SNAKE_CASE__ = ocr_lang SCREAMING_SNAKE_CASE__ = tesseract_config def lowercase_ ( self : int , __lowerCamelCase : np.ndarray , __lowerCamelCase : Dict[str, int] , __lowerCamelCase : PILImageResampling = PILImageResampling.BILINEAR , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : str , ) -> np.ndarray: SCREAMING_SNAKE_CASE__ = get_size_dict(__lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) SCREAMING_SNAKE_CASE__ = (size['''height'''], size['''width''']) return resize(__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def lowercase_ ( self : str , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[int, float] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : Optional[Any] , ) -> np.ndarray: return rescale(__lowerCamelCase , scale=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def lowercase_ ( self : Optional[int] , __lowerCamelCase : np.ndarray , __lowerCamelCase : Union[float, Iterable[float]] , __lowerCamelCase : Union[float, Iterable[float]] , __lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase : List[str] , ) -> np.ndarray: return normalize(__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def lowercase_ ( self : Tuple , __lowerCamelCase : ImageInput , __lowerCamelCase : bool = None , __lowerCamelCase : Dict[str, int] = None , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : bool = None , __lowerCamelCase : float = None , __lowerCamelCase : bool = None , __lowerCamelCase : Union[float, Iterable[float]] = None , __lowerCamelCase : Union[float, Iterable[float]] = None , __lowerCamelCase : bool = None , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[Union[str, TensorType]] = None , __lowerCamelCase : ChannelDimension = ChannelDimension.FIRST , **__lowerCamelCase : int , ) -> PIL.Image.Image: SCREAMING_SNAKE_CASE__ = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE__ = size if size is not None else self.size SCREAMING_SNAKE_CASE__ = get_size_dict(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE__ = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE__ = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE__ = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE__ = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE__ = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE__ = apply_ocr if apply_ocr is not None else self.apply_ocr SCREAMING_SNAKE_CASE__ = ocr_lang if ocr_lang is not None else self.ocr_lang SCREAMING_SNAKE_CASE__ = tesseract_config if tesseract_config is not None else self.tesseract_config SCREAMING_SNAKE_CASE__ = make_list_of_images(__lowerCamelCase ) if not valid_images(__lowerCamelCase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''If do_normalize is True, image_mean and image_std must be specified.''' ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE__ = [to_numpy_array(__lowerCamelCase ) for image in images] # Tesseract OCR to get words + normalized bounding boxes if apply_ocr: requires_backends(self , '''pytesseract''' ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] for image in images: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = apply_tesseract(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) words_batch.append(__lowerCamelCase ) boxes_batch.append(__lowerCamelCase ) if do_resize: SCREAMING_SNAKE_CASE__ = [self.resize(image=__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE__ = [self.rescale(image=__lowerCamelCase , scale=__lowerCamelCase ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE__ = [self.normalize(image=__lowerCamelCase , mean=__lowerCamelCase , std=__lowerCamelCase ) for image in images] SCREAMING_SNAKE_CASE__ = [to_channel_dimension_format(__lowerCamelCase , __lowerCamelCase ) for image in images] SCREAMING_SNAKE_CASE__ = BatchFeature(data={'''pixel_values''': images} , tensor_type=__lowerCamelCase ) if apply_ocr: SCREAMING_SNAKE_CASE__ = words_batch SCREAMING_SNAKE_CASE__ = boxes_batch return data
314
def UpperCAmelCase_ ( _A = 1_00_00_00 ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = set(range(3 , _A , 2 ) ) primes.add(2 ) for p in range(3 , _A , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , _A , _A ) ) ) SCREAMING_SNAKE_CASE__ = [float(_A ) for n in range(limit + 1 )] for p in primes: for n in range(_A , limit + 1 , _A ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"{solution() = }")
314
1
import math def A(__a: float , __a: float ): return math.pow(__a , 2 ) - a def A(__a: float ): return 2 * x def A(__a: float ): lowerCAmelCase_ = 2.0 while start <= a: lowerCAmelCase_ = math.pow(__a , 2 ) return start def A(__a: float , __a: int = 9999 , __a: float = 0.00_0000_0000_0001 ): if a < 0: raise ValueError("math domain error" ) lowerCAmelCase_ = get_initial_point(__a ) for _ in range(__a ): lowerCAmelCase_ = value lowerCAmelCase_ = value - fx(__a , __a ) / fx_derivative(__a ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
22
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase__ = { '''configuration_time_series_transformer''': [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimeSeriesTransformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimeSeriesTransformerForPrediction''', '''TimeSeriesTransformerModel''', '''TimeSeriesTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
22
1
from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """microsoft/xprophetnet-large-wiki100-cased""": ( """https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json""" ), } class UpperCAmelCase_ ( lowercase ): """simple docstring""" UpperCamelCase_ : List[str] ='xlm-prophetnet' UpperCamelCase_ : Union[str, Any] =['past_key_values'] UpperCamelCase_ : Union[str, Any] ={ 'num_attention_heads': 'num_encoder_attention_heads', } def __init__( self , SCREAMING_SNAKE_CASE_ = 0.1 , SCREAMING_SNAKE_CASE_ = "gelu" , SCREAMING_SNAKE_CASE_ = 3_0522 , SCREAMING_SNAKE_CASE_ = 1024 , SCREAMING_SNAKE_CASE_ = 4096 , SCREAMING_SNAKE_CASE_ = 12 , SCREAMING_SNAKE_CASE_ = 16 , SCREAMING_SNAKE_CASE_ = 4096 , SCREAMING_SNAKE_CASE_ = 12 , SCREAMING_SNAKE_CASE_ = 16 , SCREAMING_SNAKE_CASE_ = 0.1 , SCREAMING_SNAKE_CASE_ = 0.1 , SCREAMING_SNAKE_CASE_ = 512 , SCREAMING_SNAKE_CASE_ = 0.02 , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = 0 , SCREAMING_SNAKE_CASE_ = 2 , SCREAMING_SNAKE_CASE_ = 32 , SCREAMING_SNAKE_CASE_ = 128 , SCREAMING_SNAKE_CASE_ = False , SCREAMING_SNAKE_CASE_ = 0.0 , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = 0 , SCREAMING_SNAKE_CASE_ = 1 , SCREAMING_SNAKE_CASE_ = 2 , **SCREAMING_SNAKE_CASE_ , ) -> Dict: UpperCamelCase :Any = vocab_size UpperCamelCase :Any = hidden_size UpperCamelCase :Optional[int] = encoder_ffn_dim UpperCamelCase :Tuple = num_encoder_layers UpperCamelCase :List[Any] = num_encoder_attention_heads UpperCamelCase :Dict = decoder_ffn_dim UpperCamelCase :int = num_decoder_layers UpperCamelCase :int = num_decoder_attention_heads UpperCamelCase :Tuple = max_position_embeddings UpperCamelCase :Optional[Any] = init_std # Normal(0, this parameter) UpperCamelCase :Optional[Any] = activation_function # parameters for xlmprophetnet UpperCamelCase :Tuple = ngram UpperCamelCase :str = num_buckets UpperCamelCase :Any = relative_max_distance UpperCamelCase :Dict = disable_ngram_loss UpperCamelCase :List[Any] = eps # 3 Types of Dropout UpperCamelCase :List[Any] = attention_dropout UpperCamelCase :List[Any] = activation_dropout UpperCamelCase :Optional[int] = dropout UpperCamelCase :List[Any] = use_cache super().__init__( pad_token_id=SCREAMING_SNAKE_CASE_ , bos_token_id=SCREAMING_SNAKE_CASE_ , eos_token_id=SCREAMING_SNAKE_CASE_ , is_encoder_decoder=SCREAMING_SNAKE_CASE_ , add_cross_attention=SCREAMING_SNAKE_CASE_ , decoder_start_token_id=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ , ) @property def UpperCAmelCase ( self ) -> int: return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ ) -> List[str]: raise NotImplementedError( '''This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and''' ''' `num_decoder_layers`.''' )
259
def _A ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : str ): UpperCamelCase :Any = len(SCREAMING_SNAKE_CASE__ ) UpperCamelCase :str = len(SCREAMING_SNAKE_CASE__ ) UpperCamelCase :int = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] UpperCamelCase :List[str] = True for i in range(SCREAMING_SNAKE_CASE__ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: UpperCamelCase :List[Any] = True if a[i].islower(): UpperCamelCase :List[Any] = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
259
1
'''simple docstring''' import torch from torch import nn class lowerCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Union[str, Any]=1 , SCREAMING_SNAKE_CASE_ : Optional[int]=False ) -> str: '''simple docstring''' super().__init__() A: Any = n_token A: Optional[Any] = d_embed A: Dict = d_proj A: str = cutoffs + [n_token] A: List[Any] = [0] + self.cutoffs A: Tuple = div_val A: List[Any] = self.cutoffs[0] A: List[Any] = len(self.cutoffs ) - 1 A: Tuple = self.shortlist_size + self.n_clusters if self.n_clusters > 0: A: Any = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) A: Optional[int] = nn.Parameter(torch.zeros(self.n_clusters ) ) A: Union[str, Any] = nn.ModuleList() A: int = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(__lowerCAmelCase , __lowerCAmelCase ) ) ) else: self.out_projs.append(__lowerCAmelCase ) self.out_layers.append(nn.Linear(__lowerCAmelCase , __lowerCAmelCase ) ) else: for i in range(len(self.cutoffs ) ): A , A: Optional[int] = self.cutoff_ends[i], self.cutoff_ends[i + 1] A: List[str] = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(__lowerCAmelCase , __lowerCAmelCase ) ) ) self.out_layers.append(nn.Linear(__lowerCAmelCase , r_idx - l_idx ) ) A: str = keep_order def _snake_case ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Dict ) -> Optional[Any]: '''simple docstring''' if proj is None: A: Union[str, Any] = nn.functional.linear(__lowerCAmelCase , __lowerCAmelCase , bias=__lowerCAmelCase ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: A: Union[str, Any] = nn.functional.linear(__lowerCAmelCase , proj.t().contiguous() ) A: List[str] = nn.functional.linear(__lowerCAmelCase , __lowerCAmelCase , bias=__lowerCAmelCase ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def _snake_case ( self : Any , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : str=None , SCREAMING_SNAKE_CASE_ : List[str]=False ) -> Optional[int]: '''simple docstring''' if labels is not None: # Shift so that tokens < n predict n A: Optional[int] = hidden[..., :-1, :].contiguous() A: List[Any] = labels[..., 1:].contiguous() A: str = hidden.view(-1 , hidden.size(-1 ) ) A: Any = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError('''Input and labels should have the same size in the batch dimension.''' ) else: A: Optional[Any] = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: A: Optional[Any] = self._compute_logit(__lowerCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: A: List[Any] = labels != -1_00 A: int = torch.zeros_like(__lowerCAmelCase , dtype=hidden.dtype , device=hidden.device ) A: str = ( -nn.functional.log_softmax(__lowerCAmelCase , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: A: Any = nn.functional.log_softmax(__lowerCAmelCase , dim=-1 ) else: # construct weights and biases A , A: str = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: A , A: str = self.cutoff_ends[i], self.cutoff_ends[i + 1] A: List[str] = self.out_layers[0].weight[l_idx:r_idx] A: Tuple = self.out_layers[0].bias[l_idx:r_idx] else: A: Any = self.out_layers[i].weight A: Optional[int] = self.out_layers[i].bias if i == 0: A: List[Any] = torch.cat([weight_i, self.cluster_weight] , dim=0 ) A: List[Any] = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(__lowerCAmelCase ) biases.append(__lowerCAmelCase ) A , A , A: Tuple = weights[0], biases[0], self.out_projs[0] A: Union[str, Any] = self._compute_logit(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) A: int = nn.functional.log_softmax(__lowerCAmelCase , dim=1 ) if labels is None: A: List[str] = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: A: Optional[Any] = torch.zeros_like(__lowerCAmelCase , dtype=hidden.dtype , device=hidden.device ) A: str = 0 A: Tuple = [0] + self.cutoffs for i in range(len(__lowerCAmelCase ) - 1 ): A , A: Optional[Any] = cutoff_values[i], cutoff_values[i + 1] if labels is not None: A: Tuple = (labels >= l_idx) & (labels < r_idx) A: Union[str, Any] = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue A: Tuple = labels.index_select(0 , __lowerCAmelCase ) - l_idx A: str = head_logprob.index_select(0 , __lowerCAmelCase ) A: List[Any] = hidden.index_select(0 , __lowerCAmelCase ) else: A: Optional[int] = hidden if i == 0: if labels is not None: A: Union[str, Any] = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: A: List[Any] = head_logprob[:, : self.cutoffs[0]] else: A , A , A: Dict = weights[i], biases[i], self.out_projs[i] A: Any = self._compute_logit(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) A: Union[str, Any] = nn.functional.log_softmax(__lowerCAmelCase , dim=1 ) A: Dict = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: A: Optional[Any] = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: A: List[Any] = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i A: Dict = logprob_i if labels is not None: if (hasattr(self , '''keep_order''' ) and self.keep_order) or keep_order: out.index_copy_(0 , __lowerCAmelCase , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def _snake_case ( self : str , SCREAMING_SNAKE_CASE_ : Any ) -> Dict: '''simple docstring''' if self.n_clusters == 0: A: List[str] = self._compute_logit(__lowerCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(__lowerCAmelCase , dim=-1 ) else: # construct weights and biases A , A: str = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: A , A: str = self.cutoff_ends[i], self.cutoff_ends[i + 1] A: List[Any] = self.out_layers[0].weight[l_idx:r_idx] A: Dict = self.out_layers[0].bias[l_idx:r_idx] else: A: str = self.out_layers[i].weight A: str = self.out_layers[i].bias if i == 0: A: List[str] = torch.cat([weight_i, self.cluster_weight] , dim=0 ) A: Optional[Any] = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(__lowerCAmelCase ) biases.append(__lowerCAmelCase ) A , A , A: str = weights[0], biases[0], self.out_projs[0] A: Union[str, Any] = self._compute_logit(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) A: Any = hidden.new_empty((head_logit.size(0 ), self.n_token) ) A: Union[str, Any] = nn.functional.log_softmax(__lowerCAmelCase , dim=1 ) A: List[str] = [0] + self.cutoffs for i in range(len(__lowerCAmelCase ) - 1 ): A , A: List[str] = cutoff_values[i], cutoff_values[i + 1] if i == 0: A: List[Any] = head_logprob[:, : self.cutoffs[0]] else: A , A , A: Any = weights[i], biases[i], self.out_projs[i] A: Tuple = self._compute_logit(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) A: Any = nn.functional.log_softmax(__lowerCAmelCase , dim=1 ) A: int = head_logprob[:, -i] + tail_logprob_i A: Tuple = logprob_i return out
358
'''simple docstring''' from __future__ import annotations from typing import Any class lowerCAmelCase_ ( UpperCAmelCase_ ): '''simple docstring''' pass class lowerCAmelCase_ : '''simple docstring''' def __init__( self : List[Any] , SCREAMING_SNAKE_CASE_ : Any ) -> None: '''simple docstring''' A: Any = data A: Node | None = None def __iter__( self : Optional[int] ) -> List[str]: '''simple docstring''' A: List[str] = self A: Dict = [] while node: if node in visited: raise ContainsLoopError visited.append(SCREAMING_SNAKE_CASE_ ) yield node.data A: str = node.next_node @property def _snake_case ( self : List[str] ) -> bool: '''simple docstring''' try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": UpperCamelCase = Node(1) UpperCamelCase = Node(2) UpperCamelCase = Node(3) UpperCamelCase = Node(4) print(root_node.has_loop) # False UpperCamelCase = root_node.next_node print(root_node.has_loop) # True UpperCamelCase = Node(5) UpperCamelCase = Node(6) UpperCamelCase = Node(5) UpperCamelCase = Node(6) print(root_node.has_loop) # False UpperCamelCase = Node(1) print(root_node.has_loop) # False
334
0
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class __A: def __init__( self , _snake_case , ) -> List[str]: '''simple docstring''' __a = parent __a = 13 __a = 7 __a = 30 __a = self.seq_length + self.mem_len __a = 15 __a = True __a = True __a = 99 __a = [10, 50, 80] __a = 32 __a = 32 __a = 4 __a = 8 __a = 128 __a = 2 __a = 2 __a = None __a = 1 __a = 0 __a = 3 __a = self.vocab_size - 1 __a = 0.01 def SCREAMING_SNAKE_CASE_ ( self ) -> List[Any]: '''simple docstring''' __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' random.seed(self.seed ) tf.random.set_seed(self.seed ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> List[Any]: '''simple docstring''' __a = TFTransfoXLModel(SCREAMING_SNAKE_CASE__ ) __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a = {'input_ids': input_ids_a, 'mems': mems_a} __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> str: '''simple docstring''' __a = TFTransfoXLLMHeadModel(SCREAMING_SNAKE_CASE__ ) __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a = {'input_ids': input_ids_a, 'labels': lm_labels} __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() __a = model([input_ids_a, mems_a] ).to_tuple() __a = {'input_ids': input_ids_a, 'mems': mems_a, 'labels': lm_labels} __a = model(SCREAMING_SNAKE_CASE__ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case ) -> Dict: '''simple docstring''' __a = TFTransfoXLForSequenceClassification(SCREAMING_SNAKE_CASE__ ) __a = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self ) -> Union[str, Any]: '''simple docstring''' __a = self.prepare_config_and_inputs() (__a) = config_and_inputs __a = {'input_ids': input_ids_a} return config, inputs_dict @require_tf class __A( lowercase__ , lowercase__ , unittest.TestCase ): snake_case_ = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) snake_case_ = () if is_tf_available() else () snake_case_ = ( { '''feature-extraction''': TFTransfoXLModel, '''text-classification''': TFTransfoXLForSequenceClassification, '''text-generation''': TFTransfoXLLMHeadModel, '''zero-shot''': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) -> str: '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' __a = TFTransfoXLModelTester(self ) __a = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , d_embed=37 ) def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[int]: '''simple docstring''' self.model_tester.set_seed() __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' self.model_tester.set_seed() __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self ) -> str: '''simple docstring''' __a = self.model_tester.prepare_config_and_inputs_for_common() __a = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: __a = model_class(SCREAMING_SNAKE_CASE__ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: __a = model.get_output_embeddings() assert isinstance(SCREAMING_SNAKE_CASE__ , tf.keras.layers.Layer ) __a = model.get_bias() assert name is None else: __a = model.get_output_embeddings() assert x is None __a = model.get_bias() assert name is None def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' pass @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = TFTransfoXLModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) @unittest.skip(reason='''This model doesn\'t play well with fit() due to not returning a single loss.''' ) def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' pass @require_tf class __A( unittest.TestCase ): @unittest.skip('''Skip test until #12651 is resolved.''' ) @slow def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' __a = TFTransfoXLLMHeadModel.from_pretrained('''transfo-xl-wt103''' ) # fmt: off __a = tf.convert_to_tensor([[33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off __a = [33,1_297,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,22,1_706,17,20_098,5,3_215,21,37,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,6_224,831,16_002,2,8,603,78_967,29_546,23,803,20,25,416,5,8,232,4,277,6,1_855,4_601,3,29_546,54,8,3_609,5,57_211,49,4,1,277,18,8,1_755,15_691,3,341,25,416,693,42_573,71,17,401,94,31,17_919,2,29_546,7_873,18,1,435,23,11_011,755,5,5_167,3,7_983,98,84,2,29_546,3_267,8,3_609,4,1,4_865,1_075,2,6_087,71,6,346,8,5_854,3,29_546,824,1_400,1_868,2,19,160,2,311,8,5_496,2,20_920,17,25,15_097,3,24,24,0,33,1,1_857,2,1,1_009,4,1_109,11_739,4_762,358,5,25,245,28,1_110,3,13,1_041,4,24,603,490,2,71_477,20_098,104_447,2,20_961,1,2_604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> __a = model.generate(SCREAMING_SNAKE_CASE__ , max_length=200 , do_sample=SCREAMING_SNAKE_CASE__ ) self.assertListEqual(output_ids[0].numpy().tolist() , SCREAMING_SNAKE_CASE__ )
6
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class SCREAMING_SNAKE_CASE__ ( lowercase__ , unittest.TestCase ): snake_case__ : Optional[Any] = TextToVideoSDPipeline snake_case__ : Optional[int] = TEXT_TO_IMAGE_PARAMS snake_case__ : str = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. snake_case__ : Optional[Any] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def SCREAMING_SNAKE_CASE ( self : str ) -> Union[str, Any]: torch.manual_seed(0 ) a_ : Optional[int] = UNetaDConditionModel( block_out_channels=(3_2, 6_4, 6_4, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') , up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') , cross_attention_dim=3_2 , attention_head_dim=4 , ) a_ : int = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=SCREAMING_SNAKE_CASE__ , set_alpha_to_one=SCREAMING_SNAKE_CASE__ , ) torch.manual_seed(0 ) a_ : int = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) a_ : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='gelu' , projection_dim=5_1_2 , ) a_ : Dict = CLIPTextModel(SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) a_ : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def SCREAMING_SNAKE_CASE ( self : List[Any] , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : List[Any]=0 ) -> List[str]: if str(SCREAMING_SNAKE_CASE__ ).startswith('mps' ): a_ : Dict = torch.manual_seed(SCREAMING_SNAKE_CASE__ ) else: a_ : Optional[Any] = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(SCREAMING_SNAKE_CASE__ ) a_ : int = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'pt', } return inputs def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Tuple: a_ : List[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator a_ : Dict = self.get_dummy_components() a_ : str = TextToVideoSDPipeline(**SCREAMING_SNAKE_CASE__ ) a_ : Union[str, Any] = sd_pipe.to(SCREAMING_SNAKE_CASE__ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) a_ : Tuple = self.get_dummy_inputs(SCREAMING_SNAKE_CASE__ ) a_ : Dict = 'np' a_ : Dict = sd_pipe(**SCREAMING_SNAKE_CASE__ ).frames a_ : int = frames[0][-3:, -3:, -1] assert frames[0].shape == (6_4, 6_4, 3) a_ : Union[str, Any] = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def SCREAMING_SNAKE_CASE ( self : Any ) -> List[Any]: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=SCREAMING_SNAKE_CASE__ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def SCREAMING_SNAKE_CASE ( self : Any ) -> str: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=SCREAMING_SNAKE_CASE__ , expected_max_diff=1E-2 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def SCREAMING_SNAKE_CASE ( self : Dict ) -> Union[str, Any]: pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def SCREAMING_SNAKE_CASE ( self : str ) -> List[str]: pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def SCREAMING_SNAKE_CASE ( self : int ) -> List[Any]: pass def SCREAMING_SNAKE_CASE ( self : Dict ) -> Union[str, Any]: return super().test_progress_bar() @slow @skip_mps class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[Any]: a_ : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy' ) a_ : Any = TextToVideoSDPipeline.from_pretrained('damo-vilab/text-to-video-ms-1.7b' ) a_ : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) a_ : Optional[Any] = pipe.to('cuda' ) a_ : Any = 'Spiderman is surfing' a_ : List[Any] = torch.Generator(device='cpu' ).manual_seed(0 ) a_ : Optional[Any] = pipe(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2_5 , output_type='pt' ).frames a_ : str = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def SCREAMING_SNAKE_CASE ( self : Any ) -> Any: a_ : Dict = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy' ) a_ : Tuple = TextToVideoSDPipeline.from_pretrained('damo-vilab/text-to-video-ms-1.7b' ) a_ : Tuple = pipe.to('cuda' ) a_ : Any = 'Spiderman is surfing' a_ : List[str] = torch.Generator(device='cpu' ).manual_seed(0 ) a_ : List[Any] = pipe(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type='pt' ).frames a_ : List[str] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
32
0
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''deepmind/language-perceiver''': '''https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json''', # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class UpperCamelCase_ (lowerCamelCase_ ): __magic_name__ = '''perceiver''' def __init__( self : Dict , lowerCAmelCase_ : Dict=256 , lowerCAmelCase_ : str=1_280 , lowerCAmelCase_ : List[str]=768 , lowerCAmelCase_ : Tuple=1 , lowerCAmelCase_ : Optional[Any]=26 , lowerCAmelCase_ : Dict=8 , lowerCAmelCase_ : Any=8 , lowerCAmelCase_ : Optional[int]=None , lowerCAmelCase_ : Tuple=None , lowerCAmelCase_ : Optional[int]="kv" , lowerCAmelCase_ : str=1 , lowerCAmelCase_ : Optional[int]=1 , lowerCAmelCase_ : Union[str, Any]="gelu" , lowerCAmelCase_ : Tuple=0.1 , lowerCAmelCase_ : str=0.0_2 , lowerCAmelCase_ : str=1e-12 , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Any=262 , lowerCAmelCase_ : Any=2_048 , lowerCAmelCase_ : Optional[int]=56 , lowerCAmelCase_ : Tuple=[368, 496] , lowerCAmelCase_ : Tuple=16 , lowerCAmelCase_ : int=1_920 , lowerCAmelCase_ : Dict=16 , lowerCAmelCase_ : int=[1, 16, 224, 224] , **lowerCAmelCase_ : List[str] , ) -> Any: super().__init__(**__snake_case ) UpperCAmelCase_ : Optional[int] = num_latents UpperCAmelCase_ : List[Any] = d_latents UpperCAmelCase_ : str = d_model UpperCAmelCase_ : str = num_blocks UpperCAmelCase_ : int = num_self_attends_per_block UpperCAmelCase_ : str = num_self_attention_heads UpperCAmelCase_ : Optional[Any] = num_cross_attention_heads UpperCAmelCase_ : List[str] = qk_channels UpperCAmelCase_ : List[Any] = v_channels UpperCAmelCase_ : Dict = cross_attention_shape_for_attention UpperCAmelCase_ : List[str] = self_attention_widening_factor UpperCAmelCase_ : Any = cross_attention_widening_factor UpperCAmelCase_ : Dict = hidden_act UpperCAmelCase_ : Dict = attention_probs_dropout_prob UpperCAmelCase_ : List[Any] = initializer_range UpperCAmelCase_ : Dict = layer_norm_eps UpperCAmelCase_ : List[Any] = use_query_residual # masked language modeling attributes UpperCAmelCase_ : Any = vocab_size UpperCAmelCase_ : Dict = max_position_embeddings # image classification attributes UpperCAmelCase_ : Any = image_size # flow attributes UpperCAmelCase_ : Union[str, Any] = train_size # multimodal autoencoding attributes UpperCAmelCase_ : str = num_frames UpperCAmelCase_ : str = audio_samples_per_frame UpperCAmelCase_ : Dict = samples_per_patch UpperCAmelCase_ : List[Any] = output_shape class UpperCamelCase_ (lowerCamelCase_ ): @property def _SCREAMING_SNAKE_CASE ( self : int ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCAmelCase_ : int = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ : Any = {0: "batch", 1: "sequence"} return OrderedDict( [ ("inputs", dynamic_axis), ("attention_mask", dynamic_axis), ] ) @property def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> float: return 1e-4 def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : Optional[TensorType] = None , lowerCAmelCase_ : int = 3 , lowerCAmelCase_ : int = 40 , lowerCAmelCase_ : int = 40 , ) -> Mapping[str, Any]: # copied from `transformers.onnx.config.OnnxConfig` and slightly altered/simplified if isinstance(__snake_case , __snake_case ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase_ : Optional[Any] = compute_effective_axis_dimension( __snake_case , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase_ : Union[str, Any] = preprocessor.num_special_tokens_to_add(__snake_case ) UpperCAmelCase_ : List[str] = compute_effective_axis_dimension( __snake_case , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__snake_case ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase_ : int = [" ".join(["a"] ) * seq_length] * batch_size UpperCAmelCase_ : List[str] = dict(preprocessor(__snake_case , return_tensors=__snake_case ) ) UpperCAmelCase_ : int = inputs.pop("input_ids" ) return inputs elif isinstance(__snake_case , __snake_case ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase_ : Tuple = compute_effective_axis_dimension(__snake_case , fixed_dimension=OnnxConfig.default_fixed_batch ) UpperCAmelCase_ : Tuple = self._generate_dummy_images(__snake_case , __snake_case , __snake_case , __snake_case ) UpperCAmelCase_ : List[str] = dict(preprocessor(images=__snake_case , return_tensors=__snake_case ) ) UpperCAmelCase_ : int = inputs.pop("pixel_values" ) return inputs else: raise ValueError( "Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor." )
371
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class UpperCamelCase_ (unittest.TestCase ): @slow def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: UpperCAmelCase_ : Union[str, Any] = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = TFAutoModel.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = AutoModel.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> List[Any]: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: UpperCAmelCase_ : Dict = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : str = TFAutoModelForPreTraining.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = AutoModelForPreTraining.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[Any]: for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : List[Any] = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = TFAutoModelForCausalLM.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = TFAutoModelForCausalLM.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = AutoModelForCausalLM.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = AutoModelForCausalLM.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Union[str, Any]: for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : List[Any] = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = TFAutoModelWithLMHead.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = AutoModelWithLMHead.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Union[str, Any]: for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Dict = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = TFAutoModelForMaskedLM.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = TFAutoModelForMaskedLM.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : List[str] = AutoModelForMaskedLM.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = AutoModelForMaskedLM.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Dict: for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : int = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Optional[int] = TFAutoModelForSeqaSeqLM.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = AutoModelForSeqaSeqLM.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = AutoModelForSeqaSeqLM.from_pretrained( lowerCAmelCase_ , output_loading_info=lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: UpperCAmelCase_ : List[str] = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = TFAutoModelForSequenceClassification.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : str = AutoModelForSequenceClassification.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Tuple: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: UpperCAmelCase_ : Tuple = AutoConfig.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : Tuple = TFAutoModelForQuestionAnswering.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : int = AutoModelForQuestionAnswering.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> List[str]: UpperCAmelCase_ : str = TFAutoModelWithLMHead.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase_ ) , 14_410 ) UpperCAmelCase_ : Optional[Any] = AutoModelWithLMHead.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase_ ) , 14_410 ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase_ : List[str] = TFAutoModelWithLMHead.from_pretrained(lowerCAmelCase_ , from_pt=lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase_ ) , 14_410 ) UpperCAmelCase_ : str = AutoModelWithLMHead.from_pretrained(lowerCAmelCase_ , from_tf=lowerCAmelCase_ ) self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase_ ) , 14_410 )
253
0
'''simple docstring''' def __UpperCAmelCase ( a_: int = 10**12 ): _UpperCAmelCase : Optional[Any] = 1 _UpperCAmelCase : str = 0 _UpperCAmelCase : List[Any] = 1 _UpperCAmelCase : Optional[Any] = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(f'{solution() = }')
145
'''simple docstring''' import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class A__ ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ : Any = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : int ) -> List[Any]: """simple docstring""" _UpperCAmelCase : int = hf_hub_download( repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) _UpperCAmelCase : int = VideoClassificationPipeline(model=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , top_k=2 ) _UpperCAmelCase : int = [ example_video_filepath, "https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4", ] return video_classifier, examples def _lowerCAmelCase ( self : Optional[Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Any ) -> int: """simple docstring""" for example in examples: _UpperCAmelCase : List[str] = video_classifier(lowerCAmelCase__ ) self.assertEqual( lowerCAmelCase__ , [ {"score": ANY(lowerCAmelCase__ ), "label": ANY(lowerCAmelCase__ )}, {"score": ANY(lowerCAmelCase__ ), "label": ANY(lowerCAmelCase__ )}, ] , ) @require_torch def _lowerCAmelCase ( self : Tuple ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase : List[Any] = "hf-internal-testing/tiny-random-VideoMAEForVideoClassification" _UpperCAmelCase : Optional[int] = VideoMAEFeatureExtractor( size={"shortest_edge": 1_0} , crop_size={"height": 1_0, "width": 1_0} ) _UpperCAmelCase : List[str] = pipeline( "video-classification" , model=lowerCAmelCase__ , feature_extractor=lowerCAmelCase__ , frame_sampling_rate=4 ) _UpperCAmelCase : Tuple = hf_hub_download(repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) _UpperCAmelCase : Tuple = video_classifier(lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}] , ) _UpperCAmelCase : Any = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}], [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}], ] , ) @require_tf def _lowerCAmelCase ( self : List[str] ) -> int: """simple docstring""" pass
145
1
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets _SCREAMING_SNAKE_CASE = '''\ @inproceedings{popovic-2015-chrf, title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation", author = "Popovi{\'c}, Maja", booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation", month = sep, year = "2015", address = "Lisbon, Portugal", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W15-3049", doi = "10.18653/v1/W15-3049", pages = "392--395", } @inproceedings{popovic-2017-chrf, title = "chr{F}++: words helping character n-grams", author = "Popovi{\'c}, Maja", booktitle = "Proceedings of the Second Conference on Machine Translation", month = sep, year = "2017", address = "Copenhagen, Denmark", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/W17-4770", doi = "10.18653/v1/W17-4770", pages = "612--618", } @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' _SCREAMING_SNAKE_CASE = '''\ ChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches, and ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation that is already present in sacrebleu. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information. ''' _SCREAMING_SNAKE_CASE = ''' Produces ChrF(++) scores for hypotheses given reference translations. Args: predictions (list of str): The predicted sentences. references (list of list of str): The references. There should be one reference sub-list for each prediction sentence. char_order (int): Character n-gram order. Defaults to `6`. word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`. beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`. lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`. whitespace (bool): If `True`, include whitespaces when extracting character n-grams. eps_smoothing (bool): If `True`, applies epsilon smoothing similar to reference chrF++.py, NLTK and Moses implementations. If `False`, it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`. Returns: \'score\' (float): The chrF (chrF++) score, \'char_order\' (int): The character n-gram order, \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++, \'beta\' (int): Determine the importance of recall w.r.t precision Examples: Example 1--a simple example of calculating chrF: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, references=reference) >>> print(results) {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2} Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2) >>> print(results) {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2} Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case: >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."] >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]] >>> chrf = datasets.load_metric("chrf") >>> results = chrf.compute(predictions=prediction, ... references=reference, ... word_order=2, ... lowercase=True) >>> print(results) {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): '''simple docstring''' def _UpperCAmelCase (self ) -> List[Any]: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse('''1.4.12''' ): raise ImportWarning( '''To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n''' '''You can install it with `pip install "sacrebleu>=1.4.12"`.''' ) return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,homepage='''https://github.com/mjpost/sacreBLEU#chrf--chrf''' ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''string''' ,id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' ,id='''sequence''' ) ,id='''references''' ), } ) ,codebase_urls=['''https://github.com/mjpost/sacreBLEU#chrf--chrf'''] ,reference_urls=[ '''https://github.com/m-popovic/chrF''', ] ,) def _UpperCAmelCase (self ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase = CHRF.CHAR_ORDER ,_lowerCamelCase = CHRF.WORD_ORDER ,_lowerCamelCase = CHRF.BETA ,_lowerCamelCase = False ,_lowerCamelCase = False ,_lowerCamelCase = False ,) -> Optional[int]: '''simple docstring''' __lowercase = len(references[0] ) if any(len(_lowerCamelCase ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) __lowercase = [[refs[i] for refs in references] for i in range(_lowerCamelCase )] __lowercase = CHRF(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) __lowercase = sb_chrf.corpus_score(_lowerCamelCase ,_lowerCamelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
217
'''simple docstring''' import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging _SCREAMING_SNAKE_CASE = ['''bart.large''', '''bart.large.mnli''', '''bart.large.cnn''', '''bart_xsum/model.pt'''] _SCREAMING_SNAKE_CASE = {'''bart.large''': BartModel, '''bart.large.mnli''': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('''0.9.0'''): raise Exception('''requires fairseq >= 0.9.0''') logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = ''' Hello world! cécé herlolip''' _SCREAMING_SNAKE_CASE = [ ('''model.classification_heads.mnli.dense.weight''', '''classification_head.dense.weight'''), ('''model.classification_heads.mnli.dense.bias''', '''classification_head.dense.bias'''), ('''model.classification_heads.mnli.out_proj.weight''', '''classification_head.out_proj.weight'''), ('''model.classification_heads.mnli.out_proj.bias''', '''classification_head.out_proj.bias'''), ] def _lowerCAmelCase ( lowerCamelCase_ : int ): __lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''_float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCamelCase_ , lowerCamelCase_ ) def _lowerCAmelCase ( lowerCamelCase_ : Tuple , lowerCamelCase_ : str , lowerCamelCase_ : Union[str, Any] ): __lowercase = dct.pop(lowerCamelCase_ ) __lowercase = val def _lowerCAmelCase ( lowerCamelCase_ : Any ): __lowercase = torch.load(lowerCamelCase_ , map_location='''cpu''' ) __lowercase = torch.hub.load('''pytorch/fairseq''' , '''bart.large.cnn''' ).eval() hub_interface.model.load_state_dict(sd['''model'''] ) return hub_interface def _lowerCAmelCase ( lowerCamelCase_ : List[str] ): __lowercase , __lowercase = emb.weight.shape __lowercase = nn.Linear(lowerCamelCase_ , lowerCamelCase_ , bias=lowerCamelCase_ ) __lowercase = emb.weight.data return lin_layer @torch.no_grad() def _lowerCAmelCase ( lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : List[Any]=None ): if not os.path.exists(lowerCamelCase_ ): __lowercase = torch.hub.load('''pytorch/fairseq''' , lowerCamelCase_ ).eval() else: __lowercase = load_xsum_checkpoint(lowerCamelCase_ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowercase = checkpoint_path.replace('''.''' , '''-''' ) __lowercase = BartConfig.from_pretrained(lowerCamelCase_ ) __lowercase = bart.encode(lowerCamelCase_ ).unsqueeze(0 ) __lowercase = BartTokenizer.from_pretrained(lowerCamelCase_ ).encode(lowerCamelCase_ , return_tensors='''pt''' ).unsqueeze(0 ) if not torch.eq(lowerCamelCase_ , lowerCamelCase_ ).all(): raise ValueError( f"converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}" ) if checkpoint_path == "bart.large.mnli": __lowercase = bart.state_dict() remove_ignore_keys_(lowerCamelCase_ ) __lowercase = state_dict['''model.decoder.embed_tokens.weight'''] for src, dest in mnli_rename_keys: rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) __lowercase = BartForSequenceClassification(lowerCamelCase_ ).eval() model.load_state_dict(lowerCamelCase_ ) __lowercase = bart.predict('''mnli''' , lowerCamelCase_ , return_logits=lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ )[0] # logits else: # no classification heads to worry about __lowercase = bart.model.state_dict() remove_ignore_keys_(lowerCamelCase_ ) __lowercase = state_dict['''decoder.embed_tokens.weight'''] __lowercase = bart.extract_features(lowerCamelCase_ ) if hf_checkpoint_name == "facebook/bart-large": __lowercase = BartModel(lowerCamelCase_ ).eval() model.load_state_dict(lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ ).model[0] else: __lowercase = BartForConditionalGeneration(lowerCamelCase_ ).eval() # an existing summarization ckpt model.model.load_state_dict(lowerCamelCase_ ) if hasattr(lowerCamelCase_ , '''lm_head''' ): __lowercase = make_linear_from_emb(model.model.shared ) __lowercase = model.model(lowerCamelCase_ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f"`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}" ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError('''Some values in `fairseq_output` are different from `new_model_outputs`''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default=None, type=str, help='''Which huggingface architecture to use: bart-large-xsum''' ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
217
1
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class a : """simple docstring""" __UpperCAmelCase : torch.Tensor # [batch_size x 3] __UpperCAmelCase : torch.Tensor # [batch_size x 3] __UpperCAmelCase : torch.Tensor # [batch_size x 3] __UpperCAmelCase : torch.Tensor # [batch_size x 3] __UpperCAmelCase : int __UpperCAmelCase : int __UpperCAmelCase : float __UpperCAmelCase : float __UpperCAmelCase : Tuple[int] def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def __snake_case ( self : List[Any] ) -> Optional[Any]: return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def __snake_case ( self : str ) -> List[str]: return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def __snake_case ( self : int ) -> torch.Tensor: __snake_case : Dict = torch.arange(self.height * self.width ) __snake_case : List[str] = torch.stack( [ pixel_indices % self.width, torch.div(lowerCamelCase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def __snake_case ( self : Optional[Any] ) -> Any: __snake_case , *__snake_case : Tuple = self.shape __snake_case : Any = int(np.prod(lowerCamelCase ) ) __snake_case : List[str] = self.get_image_coords() __snake_case : List[Any] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __snake_case : Any = self.get_camera_rays(lowerCamelCase ) __snake_case : Union[str, Any] = rays.view(lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case , *__snake_case , __snake_case : Any = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __snake_case : List[str] = coords.view(lowerCamelCase , -1 , 2 ) __snake_case : Dict = self.resolution() __snake_case : Optional[int] = self.fov() __snake_case : Dict = (flat.float() / (res - 1)) * 2 - 1 __snake_case : Dict = fracs * torch.tan(fov / 2 ) __snake_case : Any = fracs.view(lowerCamelCase , -1 , 2 ) __snake_case : Dict = ( self.z.view(lowerCamelCase , 1 , 3 ) + self.x.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) __snake_case : Optional[int] = directions / directions.norm(dim=-1 , keepdim=lowerCamelCase ) __snake_case : List[str] = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCamelCase , *lowerCamelCase , 2 , 3 ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> "DifferentiableProjectiveCamera": assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCamelCase , height=lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Any = [] __snake_case : Optional[int] = [] __snake_case : Any = [] __snake_case : int = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __snake_case : str = np.array([np.sin(__lowerCamelCase ), np.cos(__lowerCamelCase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __snake_case : int = -z * 4 __snake_case : Optional[int] = np.array([np.cos(__lowerCamelCase ), -np.sin(__lowerCamelCase ), 0.0] ) __snake_case : Tuple = np.cross(__lowerCamelCase , __lowerCamelCase ) origins.append(__lowerCamelCase ) xs.append(__lowerCamelCase ) ys.append(__lowerCamelCase ) zs.append(__lowerCamelCase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(__lowerCamelCase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(__lowerCamelCase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(__lowerCamelCase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(__lowerCamelCase , axis=0 ) ).float() , width=__lowerCamelCase , height=__lowerCamelCase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(__lowerCamelCase )) , )
123
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _snake_case : int = logging.get_logger(__name__) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : str = None , lowerCamelCase : uuid.UUID = None , lowerCamelCase : Dict=None , lowerCamelCase : Union[str, Any]=None ) -> int: if not conversation_id: __snake_case : Optional[Any] = uuid.uuida() if past_user_inputs is None: __snake_case : List[Any] = [] if generated_responses is None: __snake_case : Optional[int] = [] __snake_case : uuid.UUID = conversation_id __snake_case : List[str] = past_user_inputs __snake_case : List[str] = generated_responses __snake_case : Optional[str] = text def __eq__( self : int , lowerCamelCase : List[str] ) -> Dict: if not isinstance(lowerCamelCase , lowerCamelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __snake_case ( self : Optional[int] , lowerCamelCase : str , lowerCamelCase : bool = False ) -> Optional[Any]: if self.new_user_input: if overwrite: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ' F'with: "{text}".' ) __snake_case : Dict = text else: logger.warning( F'User input added while unprocessed input was existing: "{self.new_user_input}" new input ' F'ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input' ) else: __snake_case : List[str] = text def __snake_case ( self : List[str] ) -> Dict: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) __snake_case : Optional[Any] = None def __snake_case ( self : List[str] , lowerCamelCase : str ) -> List[Any]: self.generated_responses.append(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Optional[Any] ) -> Dict: __snake_case : Any = F'Conversation id: {self.uuid} \n' for is_user, text in self.iter_texts(): __snake_case : List[Any] = "user" if is_user else "bot" output += F'{name} >> {text} \n' return output @add_end_docstrings( _lowerCAmelCase , R"\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n " , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : Optional[Any] , **lowerCamelCase : List[str] ) -> Any: super().__init__(*lowerCamelCase , **lowerCamelCase ) if self.tokenizer.pad_token_id is None: __snake_case : Dict = self.tokenizer.eos_token def __snake_case ( self : Dict , lowerCamelCase : List[str]=None , lowerCamelCase : int=None , lowerCamelCase : Optional[Any]=None , **lowerCamelCase : Any ) -> Any: __snake_case : Union[str, Any] = {} __snake_case : Optional[int] = {} __snake_case : Optional[Any] = {} if min_length_for_response is not None: __snake_case : int = min_length_for_response if minimum_tokens is not None: __snake_case : Tuple = minimum_tokens if "max_length" in generate_kwargs: __snake_case : Any = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: __snake_case : Any = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowerCamelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self : List[str] , lowerCamelCase : Union[Conversation, List[Conversation]] , lowerCamelCase : Optional[Any]=0 , **lowerCamelCase : Optional[Any] ) -> Union[str, Any]: __snake_case : Optional[Any] = super().__call__(lowerCamelCase , num_workers=lowerCamelCase , **lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ) and len(lowerCamelCase ) == 1: return outputs[0] return outputs def __snake_case ( self : Any , lowerCamelCase : Conversation , lowerCamelCase : Any=32 ) -> Dict[str, Any]: if not isinstance(lowerCamelCase , lowerCamelCase ): raise ValueError("ConversationalPipeline, expects Conversation as inputs" ) if conversation.new_user_input is None: raise ValueError( F'Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ' "Add user inputs with the conversation's `add_user_input` method" ) if hasattr(self.tokenizer , "_build_conversation_input_ids" ): __snake_case : Tuple = self.tokenizer._build_conversation_input_ids(lowerCamelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version __snake_case : Tuple = self._legacy_parse_and_tokenize(lowerCamelCase ) if self.framework == "pt": __snake_case : Union[str, Any] = torch.LongTensor([input_ids] ) elif self.framework == "tf": __snake_case : Union[str, Any] = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def __snake_case ( self : List[str] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict=10 , **lowerCamelCase : Tuple ) -> List[str]: __snake_case : Tuple = generate_kwargs.get("max_length" , self.model.config.max_length ) __snake_case : List[str] = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(F'Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})' ) __snake_case : Optional[int] = max_length - minimum_tokens __snake_case : List[Any] = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: __snake_case : str = model_inputs["attention_mask"][:, -trim:] __snake_case : Any = model_inputs.pop("conversation" ) __snake_case : str = max_length __snake_case : Optional[int] = self.model.generate(**lowerCamelCase , **lowerCamelCase ) if self.model.config.is_encoder_decoder: __snake_case : List[Any] = 1 else: __snake_case : Union[str, Any] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __snake_case ( self : Any , lowerCamelCase : List[Any] , lowerCamelCase : Tuple=True ) -> Any: __snake_case : Optional[int] = model_outputs["output_ids"] __snake_case : Optional[Any] = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , ) __snake_case : Optional[int] = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowerCamelCase ) return conversation def __snake_case ( self : Optional[Any] , lowerCamelCase : Conversation ) -> Dict: __snake_case : Optional[Any] = self.tokenizer.eos_token_id __snake_case : Any = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) ) if len(lowerCamelCase ) > self.tokenizer.model_max_length: __snake_case : Tuple = input_ids[-self.tokenizer.model_max_length :] return input_ids
123
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def A ( a_=None ) -> Optional[int]: if subparsers is not None: __UpperCamelCase : Optional[Any] =subparsers.add_parser('test' ) else: __UpperCamelCase : str =argparse.ArgumentParser('Accelerate test command' ) parser.add_argument( '--config_file' ,default=lowercase_ ,help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) ,) if subparsers is not None: parser.set_defaults(func=lowercase_ ) return parser def A ( a_ ) -> int: __UpperCamelCase : Optional[int] =os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['test_utils', 'scripts', 'test_script.py'] ) if args.config_file is None: __UpperCamelCase : List[str] =script_name else: __UpperCamelCase : str =F'--config_file={args.config_file} {script_name}' __UpperCamelCase : Optional[int] =['accelerate-launch'] + test_args.split() __UpperCamelCase : int =execute_subprocess_async(lowercase_ ,env=os.environ.copy() ) if result.returncode == 0: print('Test is a success! You are ready for your distributed training!' ) def A ( ) -> List[str]: __UpperCamelCase : str =test_command_parser() __UpperCamelCase : Union[str, Any] =parser.parse_args() test_command(lowercase_ ) if __name__ == "__main__": main()
362
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Optional[int]: # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCamelCase : Optional[int] =TapasConfig.from_json_file(a_ ) # set absolute/relative position embeddings parameter __UpperCamelCase : str =reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCamelCase : Optional[Any] =TapasForQuestionAnswering(config=a_ ) elif task == "WTQ": # run_task_main.py hparams __UpperCamelCase : Optional[int] =4 __UpperCamelCase : Optional[Any] =True # hparam_utils.py hparams __UpperCamelCase : int =0.664_694 __UpperCamelCase : Any =0.207_951 __UpperCamelCase : Tuple =0.121_194 __UpperCamelCase : List[str] =True __UpperCamelCase : Dict =True __UpperCamelCase : Optional[Any] =False __UpperCamelCase : Optional[int] =0.0_352_513 __UpperCamelCase : Optional[Any] =TapasForQuestionAnswering(config=a_ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCamelCase : List[Any] =4 __UpperCamelCase : List[str] =False # hparam_utils.py hparams __UpperCamelCase : List[str] =36.4_519 __UpperCamelCase : Dict =0.903_421 __UpperCamelCase : List[Any] =222.088 __UpperCamelCase : Optional[Any] =True __UpperCamelCase : Optional[int] =True __UpperCamelCase : Dict =True __UpperCamelCase : Dict =0.763_141 __UpperCamelCase : Union[str, Any] =TapasForQuestionAnswering(config=a_ ) elif task == "TABFACT": __UpperCamelCase : List[Any] =TapasForSequenceClassification(config=a_ ) elif task == "MLM": __UpperCamelCase : Optional[Any] =TapasForMaskedLM(config=a_ ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCamelCase : Optional[Any] =TapasModel(config=a_ ) else: raise ValueError(F'Task {task} not supported.' ) print(F'Building PyTorch model from configuration: {config}' ) # Load weights from tf checkpoint load_tf_weights_in_tapas(a_ ,a_ ,a_ ) # Save pytorch-model (weights and configuration) print(F'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(a_ ) # Save tokenizer files print(F'Save tokenizer files to {pytorch_dump_path}' ) __UpperCamelCase : Optional[Any] =TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' ,model_max_length=512 ) tokenizer.save_pretrained(a_ ) print('Used relative position embeddings:' ,model.config.reset_position_index_per_cell ) if __name__ == "__main__": A_ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''SQA''', type=str, help='''Model task for which to convert a checkpoint. Defaults to SQA.''' ) parser.add_argument( '''--reset_position_index_per_cell''', default=False, action='''store_true''', help='''Whether to use relative position embeddings or not. Defaults to True.''', ) parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--tapas_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained TAPAS model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A_ :Union[str, Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
245
0
'''simple docstring''' import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging _UpperCamelCase = logging.get_logger(__name__) def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase=False ) -> Tuple: try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise if not is_sharded: __lowerCamelCase : str = os.path.abspath(_lowerCAmelCase ) logger.info(F'Loading PyTorch weights from {pt_path}' ) __lowerCamelCase : Optional[Any] = torch.load(_lowerCAmelCase ,map_location='cpu' ) logger.info(F'PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.' ) __lowerCamelCase : Tuple = convert_pytorch_state_dict_to_flax(_lowerCAmelCase ,_lowerCAmelCase ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files __lowerCamelCase : Union[str, Any] = convert_pytorch_sharded_state_dict_to_flax(_lowerCAmelCase ,_lowerCAmelCase ) return flax_state_dict def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,) -> (Tuple[str], np.ndarray): def is_key_or_prefix_key_in_dict(_lowerCAmelCase ) -> bool: return len(set(_lowerCAmelCase ) & {key, (model_prefix,) + key} ) > 0 # layer norm __lowerCamelCase : List[str] = pt_tuple_key[:-1] + ('scale',) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(_lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean __lowerCamelCase : int = pt_tuple_key[:-1] + ('mean',) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(_lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var __lowerCamelCase : List[str] = pt_tuple_key[:-1] + ('var',) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(_lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # embedding __lowerCamelCase : Tuple = pt_tuple_key[:-1] + ('embedding',) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(_lowerCAmelCase ): return renamed_pt_tuple_key, pt_tensor # conv layer __lowerCamelCase : List[str] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(_lowerCAmelCase ): __lowerCamelCase : List[Any] = pt_tensor.transpose(2 ,3 ,1 ,0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __lowerCamelCase : Optional[Any] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(_lowerCAmelCase ): __lowerCamelCase : Dict = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __lowerCamelCase : Tuple = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __lowerCamelCase : int = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 __lowerCamelCase : Optional[int] = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): __lowerCamelCase : Union[str, Any] = pt_tuple_key[-2] + '_g' elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): __lowerCamelCase : Tuple = pt_tuple_key[-2] + '_v' if name is not None: __lowerCamelCase : Any = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def a_ ( _lowerCAmelCase ,_lowerCAmelCase ) -> Optional[int]: # convert pytorch tensor to numpy __lowerCamelCase : Optional[int] = {k: v.numpy() for k, v in pt_state_dict.items()} __lowerCamelCase : str = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: __lowerCamelCase : List[Any] = flax_model.params['params'] else: __lowerCamelCase : List[str] = flax_model.params __lowerCamelCase : Tuple = flatten_dict(_lowerCAmelCase ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __lowerCamelCase : Optional[int] = flatten_dict(flax_model.params['batch_stats'] ) random_flax_state_dict.update(_lowerCAmelCase ) __lowerCamelCase : str = {} __lowerCamelCase : Union[str, Any] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) __lowerCamelCase : List[str] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __lowerCamelCase : Any = tuple(pt_key.split('.' ) ) # remove base model prefix if necessary __lowerCamelCase : Any = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __lowerCamelCase : Optional[Any] = pt_tuple_key[1:] # Correctly rename weight parameters __lowerCamelCase ,__lowerCamelCase : Dict = rename_key_and_reshape_tensor( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) # add model prefix if necessary __lowerCamelCase : Dict = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __lowerCamelCase : Dict = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: __lowerCamelCase : List[Any] = jnp.asarray(_lowerCAmelCase ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_lowerCAmelCase ,_lowerCAmelCase ) continue # also add unexpected weight so that warning is thrown __lowerCamelCase : List[Any] = jnp.asarray(_lowerCAmelCase ) else: # also add unexpected weight so that warning is thrown __lowerCamelCase : str = jnp.asarray(_lowerCAmelCase ) return unflatten_dict(_lowerCAmelCase ) def a_ ( _lowerCAmelCase ,_lowerCAmelCase ) -> Any: import torch # Load the index __lowerCamelCase : Optional[int] = {} for shard_file in shard_filenames: # load using msgpack utils __lowerCamelCase : Dict = torch.load(_lowerCAmelCase ) __lowerCamelCase : Dict = {k: v.numpy() for k, v in pt_state_dict.items()} __lowerCamelCase : Optional[Any] = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __lowerCamelCase : str = flax_model.params['params'] __lowerCamelCase : Tuple = flatten_dict(_lowerCAmelCase ) random_flax_state_dict.update(flatten_dict(flax_model.params['batch_stats'] ) ) else: __lowerCamelCase : Dict = flax_model.params __lowerCamelCase : Optional[int] = flatten_dict(_lowerCAmelCase ) __lowerCamelCase : List[Any] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) __lowerCamelCase : List[str] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __lowerCamelCase : Optional[Any] = tuple(pt_key.split('.' ) ) # remove base model prefix if necessary __lowerCamelCase : List[Any] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __lowerCamelCase : Optional[int] = pt_tuple_key[1:] # Correctly rename weight parameters __lowerCamelCase ,__lowerCamelCase : Any = rename_key_and_reshape_tensor( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ) # add model prefix if necessary __lowerCamelCase : Tuple = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __lowerCamelCase : Optional[Any] = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: __lowerCamelCase : int = jnp.asarray(_lowerCAmelCase ) continue if "var" in flax_key[-1]: __lowerCamelCase : Tuple = jnp.asarray(_lowerCAmelCase ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_lowerCAmelCase ,_lowerCAmelCase ) continue # also add unexpected weight so that warning is thrown __lowerCamelCase : Optional[int] = jnp.asarray(_lowerCAmelCase ) else: # also add unexpected weight so that warning is thrown __lowerCamelCase : Tuple = jnp.asarray(_lowerCAmelCase ) return unflatten_dict(_lowerCAmelCase ) def a_ ( _lowerCAmelCase ,_lowerCAmelCase ) -> Union[str, Any]: __lowerCamelCase : Union[str, Any] = os.path.abspath(_lowerCAmelCase ) logger.info(F'Loading Flax weights from {flax_checkpoint_path}' ) # import correct flax class __lowerCamelCase : str = getattr(_lowerCAmelCase ,'Flax' + model.__class__.__name__ ) # load flax weight dict with open(_lowerCAmelCase ,'rb' ) as state_f: try: __lowerCamelCase : Tuple = from_bytes(_lowerCAmelCase ,state_f.read() ) except UnpicklingError: raise EnvironmentError(F'Unable to convert {flax_checkpoint_path} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(_lowerCAmelCase ,_lowerCAmelCase ) def a_ ( _lowerCAmelCase ,_lowerCAmelCase ) -> Optional[Any]: try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights __lowerCamelCase : Any = flatten_dict(jax.tree_util.tree_map(lambda _lowerCAmelCase : x.dtype == jnp.bfloataa ,_lowerCAmelCase ) ).values() if any(_lowerCAmelCase ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) __lowerCamelCase : Dict = jax.tree_util.tree_map( lambda _lowerCAmelCase : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params ,_lowerCAmelCase ) __lowerCamelCase : Any = flatten_dict(_lowerCAmelCase ) __lowerCamelCase : Union[str, Any] = pt_model.state_dict() __lowerCamelCase : Union[str, Any] = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split('.' )[0] for k in pt_model_dict.keys()} ) __lowerCamelCase : Tuple = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split('.' )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys __lowerCamelCase : Any = [] __lowerCamelCase : Union[str, Any] = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): __lowerCamelCase : List[str] = flax_key_tuple[0] == pt_model.base_model_prefix __lowerCamelCase : Dict = '.'.join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: __lowerCamelCase : List[Any] = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: __lowerCamelCase : Optional[Any] = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(_lowerCAmelCase ) not in pt_model_dict: # conv layer __lowerCamelCase : Tuple = flax_key_tuple[:-1] + ('weight',) __lowerCamelCase : Tuple = jnp.transpose(_lowerCAmelCase ,(3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(_lowerCAmelCase ) not in pt_model_dict: # linear layer __lowerCamelCase : Dict = flax_key_tuple[:-1] + ('weight',) __lowerCamelCase : str = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: __lowerCamelCase : Optional[Any] = flax_key_tuple[:-1] + ('weight',) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: __lowerCamelCase : str = flax_key_tuple[:-1] + ('running_mean',) elif "var" in flax_key_tuple[-1]: __lowerCamelCase : int = flax_key_tuple[:-1] + ('running_var',) if "batch_stats" in flax_state: __lowerCamelCase : Tuple = '.'.join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: __lowerCamelCase : Optional[int] = '.'.join(_lowerCAmelCase ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. __lowerCamelCase : Tuple = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: __lowerCamelCase : str = key.split('.' ) __lowerCamelCase : Tuple = None if key_components[-3::2] == ["parametrizations", "original0"]: __lowerCamelCase : Any = key_components[-2] + '_g' elif key_components[-3::2] == ["parametrizations", "original1"]: __lowerCamelCase : Tuple = key_components[-2] + '_v' if name is not None: __lowerCamelCase : Optional[int] = key_components[:-3] + [name] __lowerCamelCase : Union[str, Any] = '.'.join(_lowerCAmelCase ) __lowerCamelCase : Optional[int] = key if flax_key in special_pt_names: __lowerCamelCase : int = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' F'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict __lowerCamelCase : Tuple = np.asarray(_lowerCAmelCase ) if not isinstance(_lowerCAmelCase ,np.ndarray ) else flax_tensor __lowerCamelCase : List[str] = torch.from_numpy(_lowerCAmelCase ) # remove from missing keys missing_keys.remove(_lowerCAmelCase ) else: # weight is not expected by PyTorch model unexpected_keys.append(_lowerCAmelCase ) pt_model.load_state_dict(_lowerCAmelCase ) # re-transform missing_keys to list __lowerCamelCase : int = list(_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' F' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' F' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' F' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) else: logger.warning(F'All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n' ) if len(_lowerCAmelCase ) > 0: logger.warning( F'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' F' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ' use it for predictions and inference.' ) else: logger.warning( F'All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n' 'If your task is similar to the task the model of the checkpoint was trained on, ' F'you can already use {pt_model.__class__.__name__} for predictions without further training.' ) return pt_model
208
'''simple docstring''' import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class lowerCamelCase_ : """simple docstring""" def __init__( self : Tuple , _a : List[Any] , _a : Dict=2 , _a : Dict=32 , _a : int=16 , _a : str=3 , _a : Optional[int]=True , _a : List[Any]=True , _a : int=32 , _a : int=4 , _a : Optional[Any]=[0, 1, 2, 3] , _a : int=4 , _a : Union[str, Any]=37 , _a : List[str]="gelu" , _a : List[str]=0.1 , _a : List[str]=0.1 , _a : Union[str, Any]=0.02 , _a : str=3 , _a : int=[1, 384, 24, 24] , _a : Optional[Any]=True , _a : Tuple=None , ) -> Tuple: __lowerCamelCase : Dict = parent __lowerCamelCase : List[Any] = batch_size __lowerCamelCase : int = image_size __lowerCamelCase : Any = patch_size __lowerCamelCase : Tuple = num_channels __lowerCamelCase : Dict = is_training __lowerCamelCase : List[str] = use_labels __lowerCamelCase : Union[str, Any] = hidden_size __lowerCamelCase : List[Any] = num_hidden_layers __lowerCamelCase : List[Any] = backbone_out_indices __lowerCamelCase : Tuple = num_attention_heads __lowerCamelCase : Optional[Any] = intermediate_size __lowerCamelCase : Any = hidden_act __lowerCamelCase : List[str] = hidden_dropout_prob __lowerCamelCase : Tuple = attention_probs_dropout_prob __lowerCamelCase : List[str] = initializer_range __lowerCamelCase : Dict = num_labels __lowerCamelCase : List[Any] = backbone_featmap_shape __lowerCamelCase : Optional[int] = scope __lowerCamelCase : str = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) __lowerCamelCase : Union[str, Any] = (image_size // patch_size) ** 2 __lowerCamelCase : Optional[Any] = num_patches + 1 def _lowercase ( self : Dict ) -> Any: __lowerCamelCase : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowerCamelCase : List[str] = None if self.use_labels: __lowerCamelCase : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __lowerCamelCase : Dict = self.get_config() return config, pixel_values, labels def _lowercase ( self : Optional[Any] ) -> List[str]: __lowerCamelCase : Optional[Any] = { 'global_padding': 'same', 'layer_type': 'bottleneck', 'depths': [3, 4, 9], 'out_features': ['stage1', 'stage2', 'stage3'], 'embedding_dynamic_padding': True, 'hidden_sizes': [96, 192, 384, 768], 'num_groups': 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=_a , backbone_featmap_shape=self.backbone_featmap_shape , ) def _lowercase ( self : Optional[int] , _a : int , _a : str , _a : Optional[int] ) -> Optional[int]: __lowerCamelCase : Any = DPTModel(config=_a ) model.to(_a ) model.eval() __lowerCamelCase : List[str] = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[Any] , _a : Union[str, Any] , _a : Optional[Any] , _a : List[Any] ) -> List[Any]: __lowerCamelCase : Dict = self.num_labels __lowerCamelCase : List[Any] = DPTForDepthEstimation(_a ) model.to(_a ) model.eval() __lowerCamelCase : List[str] = model(_a ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def _lowercase ( self : List[str] , _a : Optional[Any] , _a : Tuple , _a : Tuple ) -> List[Any]: __lowerCamelCase : Union[str, Any] = self.num_labels __lowerCamelCase : Optional[Any] = DPTForSemanticSegmentation(_a ) model.to(_a ) model.eval() __lowerCamelCase : int = model(_a , labels=_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: __lowerCamelCase : Tuple = self.prepare_config_and_inputs() __lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase : str = config_and_inputs __lowerCamelCase : Optional[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): """simple docstring""" a_ =(DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () a_ =( { """depth-estimation""": DPTForDepthEstimation, """feature-extraction""": DPTModel, """image-segmentation""": DPTForSemanticSegmentation, } if is_torch_available() else {} ) a_ =False a_ =False a_ =False def _lowercase ( self : Dict ) -> Any: __lowerCamelCase : Optional[Any] = DPTModelTester(self ) __lowerCamelCase : Union[str, Any] = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def _lowercase ( self : Tuple ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason='DPT does not use inputs_embeds' ) def _lowercase ( self : Dict ) -> Union[str, Any]: pass def _lowercase ( self : Dict ) -> str: __lowerCamelCase ,__lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase : Union[str, Any] = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCamelCase : List[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def _lowercase ( self : List[str] ) -> Any: __lowerCamelCase ,__lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCamelCase : str = model_class(_a ) __lowerCamelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCamelCase : Optional[Any] = [*signature.parameters.keys()] __lowerCamelCase : Optional[int] = ['pixel_values'] self.assertListEqual(arg_names[:1] , _a ) def _lowercase ( self : Dict ) -> Any: __lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def _lowercase ( self : Union[str, Any] ) -> int: __lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*_a ) def _lowercase ( self : List[Any] ) -> Any: __lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_a ) def _lowercase ( self : Optional[int] ) -> Dict: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __lowerCamelCase ,__lowerCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase : Optional[int] = True if model_class in get_values(_a ): continue __lowerCamelCase : Tuple = model_class(_a ) model.to(_a ) model.train() __lowerCamelCase : str = self._prepare_for_class(_a , _a , return_labels=_a ) __lowerCamelCase : Dict = model(**_a ).loss loss.backward() def _lowercase ( self : Union[str, Any] ) -> str: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __lowerCamelCase ,__lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase : Tuple = False __lowerCamelCase : List[str] = True if model_class in get_values(_a ) or not model_class.supports_gradient_checkpointing: continue __lowerCamelCase : Optional[int] = model_class(_a ) model.to(_a ) model.gradient_checkpointing_enable() model.train() __lowerCamelCase : List[Any] = self._prepare_for_class(_a , _a , return_labels=_a ) __lowerCamelCase : str = model(**_a ).loss loss.backward() def _lowercase ( self : Dict ) -> Optional[Any]: __lowerCamelCase ,__lowerCamelCase : int = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase : Union[str, Any] = _config_zero_init(_a ) for model_class in self.all_model_classes: __lowerCamelCase : List[Any] = model_class(config=_a ) # Skip the check for the backbone __lowerCamelCase : Any = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": __lowerCamelCase : Dict = [f'{name}.{key}' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def _lowercase ( self : Dict ) -> Optional[int]: pass @slow def _lowercase ( self : Any ) -> int: for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: __lowerCamelCase : Union[str, Any] = DPTModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def _lowercase ( self : List[Any] ) -> List[Any]: # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type __lowerCamelCase ,__lowerCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() __lowerCamelCase : List[Any] = 'add' with self.assertRaises(_a ): __lowerCamelCase : int = DPTForDepthEstimation(_a ) def a_ ( ) -> str: __lowerCamelCase : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision @slow class lowerCamelCase_ ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Dict ) -> Tuple: __lowerCamelCase : Any = DPTImageProcessor.from_pretrained('Intel/dpt-hybrid-midas' ) __lowerCamelCase : Any = DPTForDepthEstimation.from_pretrained('Intel/dpt-hybrid-midas' ).to(_a ) __lowerCamelCase : Any = prepare_img() __lowerCamelCase : Dict = image_processor(images=_a , return_tensors='pt' ).to(_a ) # forward pass with torch.no_grad(): __lowerCamelCase : Any = model(**_a ) __lowerCamelCase : Any = outputs.predicted_depth # verify the predicted depth __lowerCamelCase : int = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape , _a ) __lowerCamelCase : Tuple = torch.tensor( [[[5.6437, 5.6146, 5.6511], [5.4371, 5.5649, 5.5958], [5.5215, 5.5184, 5.5293]]] ).to(_a ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , _a , atol=1e-4 ) )
208
1
import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __A = re.compile(R"\s+") def lowerCamelCase_ ( UpperCamelCase__ : List[str] ) -> str: """simple docstring""" return {"hash": hashlib.mda(re.sub(_UpperCAmelCase , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def lowerCamelCase_ ( UpperCamelCase__ : Union[str, Any] ) -> str: """simple docstring""" __lowerCamelCase = [len(_UpperCAmelCase ) for line in example["content"].splitlines()] return {"line_mean": np.mean(_UpperCAmelCase ), "line_max": max(_UpperCAmelCase )} def lowerCamelCase_ ( UpperCamelCase__ : str ) -> str: """simple docstring""" __lowerCamelCase = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : Optional[Any] ) -> Dict: """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def lowerCamelCase_ ( UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[str]=5 ) -> Optional[int]: """simple docstring""" __lowerCamelCase = ["auto-generated", "autogenerated", "automatically generated"] __lowerCamelCase = example["content"].splitlines() for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def lowerCamelCase_ ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any]=5 , UpperCamelCase__ : int=0.05 ) -> List[str]: """simple docstring""" __lowerCamelCase = ["unit tests", "test file", "configuration file"] __lowerCamelCase = example["content"].splitlines() __lowerCamelCase = 0 __lowerCamelCase = 0 # first test for _, line in zip(range(_UpperCAmelCase ) , _UpperCAmelCase ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test __lowerCamelCase = example["content"].count('\n' ) __lowerCamelCase = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def lowerCamelCase_ ( UpperCamelCase__ : str ) -> Optional[Any]: """simple docstring""" __lowerCamelCase = ["def ", "class ", "for ", "while "] __lowerCamelCase = example["content"].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : Optional[Any]=4 ) -> Optional[Any]: """simple docstring""" __lowerCamelCase = example["content"].splitlines() __lowerCamelCase = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def lowerCamelCase_ ( UpperCamelCase__ : int ) -> List[Any]: """simple docstring""" __lowerCamelCase = tokenizer(example['content'] , truncation=_UpperCAmelCase )["input_ids"] __lowerCamelCase = len(example['content'] ) / len(_UpperCAmelCase ) return {"ratio": ratio} def lowerCamelCase_ ( UpperCamelCase__ : str ) -> List[str]: """simple docstring""" __lowerCamelCase = {} results.update(get_hash(_UpperCAmelCase ) ) results.update(line_stats(_UpperCAmelCase ) ) results.update(alpha_stats(_UpperCAmelCase ) ) results.update(char_token_ratio(_UpperCAmelCase ) ) results.update(is_autogenerated(_UpperCAmelCase ) ) results.update(is_config_or_test(_UpperCAmelCase ) ) results.update(has_no_keywords(_UpperCAmelCase ) ) results.update(has_few_assignments(_UpperCAmelCase ) ) return results def lowerCamelCase_ ( UpperCamelCase__ : Dict , UpperCamelCase__ : str , UpperCamelCase__ : Optional[int] ) -> int: """simple docstring""" if not check_uniques(_UpperCAmelCase , _UpperCAmelCase ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def lowerCamelCase_ ( UpperCamelCase__ : List[Any] ) -> Optional[Any]: """simple docstring""" with open(_UpperCAmelCase , 'rb' ) as f_in: with gzip.open(str(_UpperCAmelCase ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) os.unlink(_UpperCAmelCase ) # Settings __A = HfArgumentParser(PreprocessingArguments) __A = parser.parse_args() if args.num_workers is None: __A = multiprocessing.cpu_count() __A = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __A = time.time() __A = load_dataset(args.dataset_name, split="train") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __A = time.time() __A = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __A = set(ds.unique("hash")) __A = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __A = time.time() __A = ds.filter(filter, fn_kwargs={"uniques": uniques, "args": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __A = time.time() __A = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __A = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / "duplicate_clusters.json", "w") as f: json.dump(duplicate_clusters, f) __A = output_dir / "data" data_dir.mkdir(exist_ok=True) __A = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __A = str(data_dir / f'''file-{file_number+1:012}.json''') __A = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
365
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase_ ( self ) -> List[str]: '''simple docstring''' __lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCamelCase__ ).to(lowerCamelCase__ ) __lowerCamelCase = AutoTokenizer.from_pretrained('google/mt5-small' ) __lowerCamelCase = tokenizer('Hello there' , return_tensors='pt' ).input_ids __lowerCamelCase = tokenizer('Hi I am' , return_tensors='pt' ).input_ids __lowerCamelCase = model(input_ids.to(lowerCamelCase__ ) , labels=labels.to(lowerCamelCase__ ) ).loss __lowerCamelCase = -(labels.shape[-1] * loss.item()) __lowerCamelCase = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
348
0
"""simple docstring""" import math def lowercase__ ( snake_case_ :List[str] , snake_case_ :List[str] ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(snake_case_ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError('''This should never happen''' ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. _lowercase : str = 'Enter the base and the power separated by a comma: ' _lowercase ,_lowercase : Tuple = map(int, input(prompt).split(',')) _lowercase ,_lowercase : int = map(int, input(prompt).split(',')) # We find the log of each number, using the function res(), which takes two # arguments. _lowercase : str = res(xa, ya) _lowercase : Optional[Any] = res(xa, ya) # We check for the largest number if resa > resa: print('Largest number is', xa, '^', ya) elif resa > resa: print('Largest number is', xa, '^', ya) else: print('Both are equal')
332
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class _UpperCAmelCase ( unittest.TestCase ): @slow def a ( self : str ): __UpperCAmelCase = XLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) __UpperCAmelCase = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house __UpperCAmelCase = torch.Size((1, 12, 7_68) ) # batch_size, sequence_length, embedding_vector_dim __UpperCAmelCase = torch.tensor( [[-0.0_101, 0.1_218, -0.0_803, 0.0_801, 0.1_327, 0.0_776, -0.1_215, 0.2_383, 0.3_338, 0.3_106, 0.0_300, 0.0_252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __UpperCAmelCase = model(_lowercase )['''last_hidden_state'''].detach() self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowercase , atol=1E-3 ) ) @slow def a ( self : str ): __UpperCAmelCase = XLMRobertaModel.from_pretrained('''xlm-roberta-large''' ) __UpperCAmelCase = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house __UpperCAmelCase = torch.Size((1, 12, 10_24) ) # batch_size, sequence_length, embedding_vector_dim __UpperCAmelCase = torch.tensor( [[-0.0_699, -0.0_318, 0.0_705, -0.1_241, 0.0_999, -0.0_520, 0.1_004, -0.1_838, -0.4_704, 0.1_437, 0.0_821, 0.0_126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __UpperCAmelCase = model(_lowercase )['''last_hidden_state'''].detach() self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _lowercase , atol=1E-3 ) )
332
1
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowercase_ ( _lowerCamelCase : Any): lowercase__ : Optional[Any] = image.size lowercase__ : int = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 lowercase__ : int = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"]) lowercase__ : Union[str, Any] = np.array(_lowerCamelCase).astype(np.floataa) / 255.0 lowercase__ : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2) lowercase__ : Dict = torch.from_numpy(_lowerCamelCase) return 2.0 * image - 1.0 class snake_case_ ( __A ): def __init__( self : str , lowercase_ : VQModel , lowercase_ : UNetaDModel , lowercase_ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> List[Any]: super().__init__() self.register_modules(vqvae=lowercase_ , unet=lowercase_ , scheduler=lowercase_ ) @torch.no_grad() def __call__( self : Tuple , lowercase_ : Union[torch.Tensor, PIL.Image.Image] = None , lowercase_ : Optional[int] = 1 , lowercase_ : Optional[int] = 1_00 , lowercase_ : Optional[float] = 0.0 , lowercase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowercase_ : Optional[str] = "pil" , lowercase_ : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowercase_ , PIL.Image.Image ): lowercase__ : Any = 1 elif isinstance(lowercase_ , torch.Tensor ): lowercase__ : List[Any] = image.shape[0] else: raise ValueError(F'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowercase_ )}''' ) if isinstance(lowercase_ , PIL.Image.Image ): lowercase__ : str = preprocess(lowercase_ ) lowercase__ : Union[str, Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image lowercase__ : Tuple = (batch_size, self.unet.config.in_channels // 2, height, width) lowercase__ : Any = next(self.unet.parameters() ).dtype lowercase__ : Optional[int] = randn_tensor(lowercase_ , generator=lowercase_ , device=self.device , dtype=lowercase_ ) lowercase__ : Tuple = image.to(device=self.device , dtype=lowercase_ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowercase_ , device=self.device ) lowercase__ : Optional[Any] = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler lowercase__ : List[str] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] lowercase__ : Optional[int] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) lowercase__ : List[str] = {} if accepts_eta: lowercase__ : int = eta for t in self.progress_bar(lowercase_ ): # concat latents and low resolution image in the channel dimension. lowercase__ : Optional[int] = torch.cat([latents, image] , dim=1 ) lowercase__ : Optional[int] = self.scheduler.scale_model_input(lowercase_ , lowercase_ ) # predict the noise residual lowercase__ : Tuple = self.unet(lowercase_ , lowercase_ ).sample # compute the previous noisy sample x_t -> x_t-1 lowercase__ : Optional[int] = self.scheduler.step(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample # decode the image latents with the VQVAE lowercase__ : Union[str, Any] = self.vqvae.decode(lowercase_ ).sample lowercase__ : str = torch.clamp(lowercase_ , -1.0 , 1.0 ) lowercase__ : Any = image / 2 + 0.5 lowercase__ : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowercase__ : Optional[Any] = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
353
import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask UpperCamelCase = logging.getLogger(__name__) class snake_case_ ( __A ): __A : int = "token-classification" def __init__( self : Tuple , lowercase_ : Dict ) -> List[str]: if type(lowercase_ ) == dict: lowercase__ : Dict = Namespace(**lowercase_ ) lowercase__ : str = import_module("tasks" ) try: lowercase__ : Tuple = getattr(lowercase_ , hparams.task_type ) lowercase__ : TokenClassificationTask = token_classification_task_clazz() except AttributeError: raise ValueError( F'''Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' F'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) lowercase__ : Optional[Any] = self.token_classification_task.get_labels(hparams.labels ) lowercase__ : int = CrossEntropyLoss().ignore_index super().__init__(lowercase_ , len(self.labels ) , self.mode ) def __UpperCamelCase ( self : Union[str, Any] , **lowercase_ : List[str] ) -> Any: return self.model(**lowercase_ ) def __UpperCamelCase ( self : Optional[Any] , lowercase_ : str , lowercase_ : Optional[int] ) -> Tuple: lowercase__ : int = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": lowercase__ : Tuple = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids lowercase__ : Optional[int] = self(**lowercase_ ) lowercase__ : Union[str, Any] = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def __UpperCamelCase ( self : Tuple ) -> Union[str, Any]: lowercase__ : Tuple = self.hparams for mode in ["train", "dev", "test"]: lowercase__ : Any = self._feature_file(lowercase_ ) if os.path.exists(lowercase_ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , lowercase_ ) lowercase__ : str = torch.load(lowercase_ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) lowercase__ : Optional[Any] = self.token_classification_task.read_examples_from_file(args.data_dir , lowercase_ ) lowercase__ : Dict = self.token_classification_task.convert_examples_to_features( lowercase_ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=lowercase_ , pad_on_left=bool(self.config.model_type in ["xlnet"] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , lowercase_ ) torch.save(lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Optional[Any] , lowercase_ : int , lowercase_ : int , lowercase_ : bool = False ) -> DataLoader: lowercase__ : str = self._feature_file(lowercase_ ) logger.info("Loading features from cached file %s" , lowercase_ ) lowercase__ : str = torch.load(lowercase_ ) lowercase__ : List[str] = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) lowercase__ : str = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: lowercase__ : Dict = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: lowercase__ : Dict = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) lowercase__ : List[str] = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) , batch_size=lowercase_ ) def __UpperCamelCase ( self : str , lowercase_ : Dict , lowercase_ : Tuple ) -> str: """Compute validation""" "" lowercase__ : Union[str, Any] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": lowercase__ : int = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids lowercase__ : List[Any] = self(**lowercase_ ) lowercase__ , lowercase__ : Any = outputs[:2] lowercase__ : Optional[Any] = logits.detach().cpu().numpy() lowercase__ : int = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def __UpperCamelCase ( self : Optional[int] , lowercase_ : Any ) -> List[Any]: lowercase__ : int = torch.stack([x["val_loss"] for x in outputs] ).mean() lowercase__ : Any = np.concatenate([x["pred"] for x in outputs] , axis=0 ) lowercase__ : Dict = np.argmax(lowercase_ , axis=2 ) lowercase__ : int = np.concatenate([x["target"] for x in outputs] , axis=0 ) lowercase__ : Any = dict(enumerate(self.labels ) ) lowercase__ : List[Any] = [[] for _ in range(out_label_ids.shape[0] )] lowercase__ : Dict = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) lowercase__ : Any = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(lowercase_ , lowercase_ ), "precision": precision_score(lowercase_ , lowercase_ ), "recall": recall_score(lowercase_ , lowercase_ ), "f1": fa_score(lowercase_ , lowercase_ ), } lowercase__ : List[Any] = dict(results.items() ) lowercase__ : List[str] = results return ret, preds_list, out_label_list def __UpperCamelCase ( self : Any , lowercase_ : Dict ) -> Dict: # when stable lowercase__ , lowercase__ , lowercase__ : Dict = self._eval_end(lowercase_ ) lowercase__ : Any = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def __UpperCamelCase ( self : str , lowercase_ : Tuple ) -> int: # updating to test_epoch_end instead of deprecated test_end lowercase__ , lowercase__ , lowercase__ : Dict = self._eval_end(lowercase_ ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 lowercase__ : Optional[int] = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def __UpperCamelCase ( lowercase_ : int , lowercase_ : Union[str, Any] ) -> Tuple: # Add NER specific options BaseTransformer.add_model_specific_args(lowercase_ , lowercase_ ) parser.add_argument( "--task_type" , default="NER" , type=lowercase_ , help="Task type to fine tune in training (e.g. NER, POS, etc)" ) parser.add_argument( "--max_seq_length" , default=1_28 , type=lowercase_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=lowercase_ , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=lowercase_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) UpperCamelCase = NERTransformer.add_model_specific_args(parser, os.getcwd()) UpperCamelCase = parser.parse_args() UpperCamelCase = NERTransformer(args) UpperCamelCase = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 UpperCamelCase = sorted(glob.glob(os.path.join(args.output_dir, '''checkpoint-epoch=*.ckpt'''), recursive=True)) UpperCamelCase = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
333
0
import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import OpenLlamaForCausalLM, OpenLlamaForSequenceClassification, OpenLlamaModel class UpperCAmelCase_ : '''simple docstring''' def __init__( self : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Any=13 , UpperCamelCase__ : str=7 , UpperCamelCase__ : Tuple=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : Tuple=False , UpperCamelCase__ : int=True , UpperCamelCase__ : int=99 , UpperCamelCase__ : List[Any]=32 , UpperCamelCase__ : Tuple=5 , UpperCamelCase__ : Optional[int]=4 , UpperCamelCase__ : str=37 , UpperCamelCase__ : Optional[Any]="gelu" , UpperCamelCase__ : Union[str, Any]=0.1 , UpperCamelCase__ : Union[str, Any]=0.1 , UpperCamelCase__ : str=512 , UpperCamelCase__ : Optional[Any]=16 , UpperCamelCase__ : str=2 , UpperCamelCase__ : Optional[int]=0.02 , UpperCamelCase__ : int=3 , UpperCamelCase__ : Optional[int]=4 , UpperCamelCase__ : str=None , ) -> Union[str, Any]: """simple docstring""" __magic_name__ = parent __magic_name__ = batch_size __magic_name__ = seq_length __magic_name__ = is_training __magic_name__ = use_input_mask __magic_name__ = use_token_type_ids __magic_name__ = use_labels __magic_name__ = vocab_size __magic_name__ = hidden_size __magic_name__ = num_hidden_layers __magic_name__ = num_attention_heads __magic_name__ = intermediate_size __magic_name__ = hidden_act __magic_name__ = hidden_dropout_prob __magic_name__ = attention_probs_dropout_prob __magic_name__ = max_position_embeddings __magic_name__ = type_vocab_size __magic_name__ = type_sequence_label_size __magic_name__ = initializer_range __magic_name__ = num_labels __magic_name__ = num_choices __magic_name__ = scope def _lowercase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __magic_name__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ = None if self.use_input_mask: __magic_name__ = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__ = None if self.use_token_type_ids: __magic_name__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__ = None __magic_name__ = None __magic_name__ = None if self.use_labels: __magic_name__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __magic_name__ = ids_tensor([self.batch_size] , self.num_choices ) __magic_name__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self : List[str] ) -> str: """simple docstring""" return OpenLlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase__ , initializer_range=self.initializer_range , use_stable_embedding=UpperCamelCase__ , ) def _lowercase ( self : str , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : List[Any] ) -> str: """simple docstring""" __magic_name__ = OpenLlamaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : Tuple , UpperCamelCase__ : int , UpperCamelCase__ : str , UpperCamelCase__ : Dict , UpperCamelCase__ : str , UpperCamelCase__ : int , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[str] , ) -> Optional[Any]: """simple docstring""" __magic_name__ = True __magic_name__ = OpenLlamaModel(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , encoder_attention_mask=UpperCamelCase__ , ) __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , ) __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple , ) -> Dict: """simple docstring""" __magic_name__ = OpenLlamaForCausalLM(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str , UpperCamelCase__ : Tuple , UpperCamelCase__ : List[str] , ) -> Optional[Any]: """simple docstring""" __magic_name__ = True __magic_name__ = True __magic_name__ = OpenLlamaForCausalLM(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() # first forward pass __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , encoder_attention_mask=UpperCamelCase__ , use_cache=UpperCamelCase__ , ) __magic_name__ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __magic_name__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) __magic_name__ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __magic_name__ = torch.cat([input_ids, next_tokens] , dim=-1 ) __magic_name__ = torch.cat([input_mask, next_mask] , dim=-1 ) __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , encoder_attention_mask=UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , )["""hidden_states"""][0] __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , encoder_hidden_states=UpperCamelCase__ , encoder_attention_mask=UpperCamelCase__ , past_key_values=UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , )["""hidden_states"""][0] # select random slice __magic_name__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() __magic_name__ = output_from_no_past[:, -3:, random_slice_idx].detach() __magic_name__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-3 ) ) def _lowercase ( self : Optional[Any] ) -> int: """simple docstring""" __magic_name__ = self.prepare_config_and_inputs() ( ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ) = config_and_inputs __magic_name__ = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _A , _A , _A , unittest.TestCase ): '''simple docstring''' a__ = ( (OpenLlamaModel, OpenLlamaForCausalLM, OpenLlamaForSequenceClassification) if is_torch_available() else () ) a__ = (OpenLlamaForCausalLM,) if is_torch_available() else () a__ = ( { """feature-extraction""": OpenLlamaModel, """text-classification""": OpenLlamaForSequenceClassification, """text-generation""": OpenLlamaForCausalLM, """zero-shot""": OpenLlamaForSequenceClassification, } if is_torch_available() else {} ) a__ = False a__ = False def _lowercase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" __magic_name__ = OpenLlamaModelTester(self ) __magic_name__ = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def _lowercase ( self : List[Any] ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : str ) -> List[str]: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _lowercase ( self : Union[str, Any] ) -> List[str]: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __magic_name__ = type self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _lowercase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = 3 __magic_name__ = input_dict["""input_ids"""] __magic_name__ = input_ids.ne(1 ).to(UpperCamelCase__ ) __magic_name__ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __magic_name__ = OpenLlamaForSequenceClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , labels=UpperCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _lowercase ( self : List[str] ) -> Tuple: """simple docstring""" __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = 3 __magic_name__ = """single_label_classification""" __magic_name__ = input_dict["""input_ids"""] __magic_name__ = input_ids.ne(1 ).to(UpperCamelCase__ ) __magic_name__ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __magic_name__ = OpenLlamaForSequenceClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , labels=UpperCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _lowercase ( self : str ) -> Dict: """simple docstring""" __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = 3 __magic_name__ = """multi_label_classification""" __magic_name__ = input_dict["""input_ids"""] __magic_name__ = input_ids.ne(1 ).to(UpperCamelCase__ ) __magic_name__ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __magic_name__ = OpenLlamaForSequenceClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , labels=UpperCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip("""Open-Llama buffers include complex numbers, which breaks this test""" ) def _lowercase ( self : str ) -> int: """simple docstring""" pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _lowercase ( self : Tuple , UpperCamelCase__ : List[Any] ) -> str: """simple docstring""" __magic_name__ , __magic_name__ = self.model_tester.prepare_config_and_inputs_for_common() __magic_name__ = ids_tensor([1, 10] , config.vocab_size ) __magic_name__ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __magic_name__ = OpenLlamaModel(UpperCamelCase__ ) original_model.to(UpperCamelCase__ ) original_model.eval() __magic_name__ = original_model(UpperCamelCase__ ).last_hidden_state __magic_name__ = original_model(UpperCamelCase__ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __magic_name__ = {"""type""": scaling_type, """factor""": 10.0} __magic_name__ = OpenLlamaModel(UpperCamelCase__ ) scaled_model.to(UpperCamelCase__ ) scaled_model.eval() __magic_name__ = scaled_model(UpperCamelCase__ ).last_hidden_state __magic_name__ = scaled_model(UpperCamelCase__ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-5 ) )
88
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : int= logging.get_logger(__name__) _a : Optional[Any]= { "SCUT-DLVCLab/lilt-roberta-en-base": ( "https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json" ), } class UpperCamelCase ( lowercase ): UpperCAmelCase : List[Any] = """lilt""" def __init__(self : Dict , _A : Any=3_05_22 , _A : Union[str, Any]=7_68 , _A : Any=12 , _A : Tuple=12 , _A : Optional[int]=30_72 , _A : Tuple="gelu" , _A : str=0.1 , _A : List[Any]=0.1 , _A : Union[str, Any]=5_12 , _A : Any=2 , _A : Tuple=0.02 , _A : List[str]=1E-12 , _A : Optional[int]=0 , _A : Optional[Any]="absolute" , _A : Any=None , _A : List[Any]=4 , _A : Optional[int]=10_24 , **_A : Union[str, Any] , ) -> Tuple: super().__init__(pad_token_id=_A , **_A) __snake_case : Optional[int] = vocab_size __snake_case : List[Any] = hidden_size __snake_case : Any = num_hidden_layers __snake_case : Optional[int] = num_attention_heads __snake_case : Optional[int] = hidden_act __snake_case : List[str] = intermediate_size __snake_case : Union[str, Any] = hidden_dropout_prob __snake_case : Dict = attention_probs_dropout_prob __snake_case : List[Any] = max_position_embeddings __snake_case : Dict = type_vocab_size __snake_case : List[Any] = initializer_range __snake_case : Optional[Any] = layer_norm_eps __snake_case : Optional[int] = position_embedding_type __snake_case : Any = classifier_dropout __snake_case : Optional[int] = channel_shrink_ratio __snake_case : Tuple = max_ad_position_embeddings
172
0
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class lowerCamelCase : '''simple docstring''' _snake_case : Optional[Any] = XGLMConfig _snake_case : int = {} _snake_case : Optional[Any] = """gelu""" def __init__( self , _UpperCamelCase , _UpperCamelCase=1_4 , _UpperCamelCase=7 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=9_9 , _UpperCamelCase=3_2 , _UpperCamelCase=2 , _UpperCamelCase=4 , _UpperCamelCase=3_7 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=5_1_2 , _UpperCamelCase=0.02 , ) -> Optional[Any]: UpperCAmelCase_ : Any = parent UpperCAmelCase_ : Tuple = batch_size UpperCAmelCase_ : Optional[Any] = seq_length UpperCAmelCase_ : Tuple = is_training UpperCAmelCase_ : List[str] = use_input_mask UpperCAmelCase_ : Union[str, Any] = use_labels UpperCAmelCase_ : Union[str, Any] = vocab_size UpperCAmelCase_ : str = d_model UpperCAmelCase_ : Dict = num_hidden_layers UpperCAmelCase_ : str = num_attention_heads UpperCAmelCase_ : Any = ffn_dim UpperCAmelCase_ : str = activation_function UpperCAmelCase_ : Optional[int] = activation_dropout UpperCAmelCase_ : str = attention_dropout UpperCAmelCase_ : Dict = max_position_embeddings UpperCAmelCase_ : Optional[int] = initializer_range UpperCAmelCase_ : Optional[int] = None UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : Optional[int] = 2 UpperCAmelCase_ : List[Any] = 1 def __UpperCAmelCase ( self ) -> int: return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def __UpperCAmelCase ( self ) -> Union[str, Any]: UpperCAmelCase_ : str = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) UpperCAmelCase_ : Union[str, Any] = None if self.use_input_mask: UpperCAmelCase_ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : Union[str, Any] = self.get_config() UpperCAmelCase_ : Optional[int] = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def __UpperCAmelCase ( self ) -> List[str]: return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=__lowercase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=__lowercase , ) def __UpperCAmelCase ( self ) -> Optional[Any]: UpperCAmelCase_ : Tuple = self.prepare_config_and_inputs() ( UpperCAmelCase_ ) : Optional[int] = config_and_inputs UpperCAmelCase_ : Optional[Any] = { '''input_ids''': input_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_tf class lowerCamelCase (__lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _snake_case : str = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () _snake_case : List[Any] = (TFXGLMForCausalLM,) if is_tf_available() else () _snake_case : Optional[int] = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) _snake_case : Optional[int] = False _snake_case : Optional[Any] = False _snake_case : str = False def __UpperCAmelCase ( self ) -> Dict: UpperCAmelCase_ : List[Any] = TFXGLMModelTester(self ) UpperCAmelCase_ : Union[str, Any] = ConfigTester(self , config_class=__lowercase , n_embd=3_7 ) def __UpperCAmelCase ( self ) -> int: self.config_tester.run_common_tests() @slow def __UpperCAmelCase ( self ) -> List[Any]: for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Optional[Any] = TFXGLMModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def __UpperCAmelCase ( self ) -> List[str]: super().test_resize_token_embeddings() @require_tf class lowerCamelCase (unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self , _UpperCamelCase=True ) -> List[Any]: UpperCAmelCase_ : List[Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) UpperCAmelCase_ : Any = tf.convert_to_tensor([[2, 2_6_8, 9_8_6_5]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off UpperCAmelCase_ : Any = [2, 2_6_8, 9_8_6_5, 6_7, 1_1, 1_9_8_8, 5_7_2_5_2, 9_8_6_5, 5, 9_8_4, 6_7, 1_9_8_8, 2_1_3_8_3_8, 1_6_5_8, 5_3, 7_0_4_4_6, 3_3, 6_6_5_7, 2_7_8, 1_5_8_1] # fmt: on UpperCAmelCase_ : Union[str, Any] = model.generate(__lowercase , do_sample=__lowercase , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , __lowercase ) @slow def __UpperCAmelCase ( self ) -> Any: UpperCAmelCase_ : Optional[int] = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) UpperCAmelCase_ : str = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) UpperCAmelCase_ : Optional[Any] = tokenizer('Today is a nice day and' , return_tensors='tf' ) UpperCAmelCase_ : Union[str, Any] = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): UpperCAmelCase_ : List[str] = model.generate(__lowercase , do_sample=__lowercase , seed=[7, 0] ) UpperCAmelCase_ : List[str] = tokenizer.decode(output_ids[0] , skip_special_tokens=__lowercase ) UpperCAmelCase_ : int = ( '''Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due''' ) self.assertEqual(__lowercase , __lowercase ) @slow def __UpperCAmelCase ( self ) -> int: UpperCAmelCase_ : Optional[Any] = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) UpperCAmelCase_ : Dict = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) UpperCAmelCase_ : Optional[int] = '''left''' # use different length sentences to test batching UpperCAmelCase_ : List[str] = [ '''This is an extremelly long sentence that only exists to test the ability of the model to cope with ''' '''left-padding, such as in batched generation. The output for the sequence below should be the same ''' '''regardless of whether left padding is applied or not. When''', '''Hello, my dog is a little''', ] UpperCAmelCase_ : Optional[Any] = tokenizer(__lowercase , return_tensors='tf' , padding=__lowercase ) UpperCAmelCase_ : Any = inputs['''input_ids'''] UpperCAmelCase_ : List[Any] = model.generate(input_ids=__lowercase , attention_mask=inputs['attention_mask'] , max_new_tokens=1_2 ) UpperCAmelCase_ : Tuple = tokenizer(sentences[0] , return_tensors='tf' ).input_ids UpperCAmelCase_ : Optional[Any] = model.generate(input_ids=__lowercase , max_new_tokens=1_2 ) UpperCAmelCase_ : Any = tokenizer(sentences[1] , return_tensors='tf' ).input_ids UpperCAmelCase_ : Any = model.generate(input_ids=__lowercase , max_new_tokens=1_2 ) UpperCAmelCase_ : Optional[Any] = tokenizer.batch_decode(__lowercase , skip_special_tokens=__lowercase ) UpperCAmelCase_ : Union[str, Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__lowercase ) UpperCAmelCase_ : List[str] = tokenizer.decode(output_padded[0] , skip_special_tokens=__lowercase ) UpperCAmelCase_ : Tuple = [ '''This is an extremelly long sentence that only exists to test the ability of the model to cope with ''' '''left-padding, such as in batched generation. The output for the sequence below should be the same ''' '''regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ''' '''a single''', '''Hello, my dog is a little bit of a shy one, but he is very friendly''', ] self.assertListEqual(__lowercase , __lowercase ) self.assertListEqual(__lowercase , [non_padded_sentence, padded_sentence] )
350
from __future__ import annotations def lowercase__ ( __snake_case : list[int] , __snake_case : int ): '''simple docstring''' if len(__snake_case ) < k or k < 0: raise ValueError('Invalid Input' ) UpperCAmelCase_ : int = sum(array[:k] ) for i in range(len(__snake_case ) - k ): UpperCAmelCase_ : List[Any] = current_sum - array[i] + array[i + k] UpperCAmelCase_ : List[Any] = max(__snake_case , __snake_case ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() __UpperCAmelCase = [randint(-1000, 1000) for i in range(100)] __UpperCAmelCase = randint(0, 110) print(F'The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}')
145
0
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE :Optional[int] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE :str = { '''facebook/encodec_24khz''': '''https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json''', '''facebook/encodec_48khz''': '''https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json''', } class A_ ( lowerCAmelCase_ ): _lowerCamelCase : Tuple = """encodec""" def __init__( self : Tuple , snake_case_ : Optional[int]=[1.5, 3.0, 6.0, 1_2.0, 2_4.0] , snake_case_ : Union[str, Any]=2_4_0_0_0 , snake_case_ : List[Any]=1 , snake_case_ : Optional[int]=False , snake_case_ : Any=None , snake_case_ : Dict=None , snake_case_ : str=1_2_8 , snake_case_ : Union[str, Any]=3_2 , snake_case_ : Optional[int]=1 , snake_case_ : Optional[int]=[8, 5, 4, 2] , snake_case_ : Dict="weight_norm" , snake_case_ : List[str]=7 , snake_case_ : Optional[Any]=7 , snake_case_ : Dict=3 , snake_case_ : Dict=2 , snake_case_ : Tuple=True , snake_case_ : Tuple="reflect" , snake_case_ : Tuple=2 , snake_case_ : Tuple=2 , snake_case_ : Union[str, Any]=1.0 , snake_case_ : List[str]=1_0_2_4 , snake_case_ : Union[str, Any]=None , snake_case_ : List[Any]=True , **snake_case_ : List[Any] , ): _UpperCAmelCase = target_bandwidths _UpperCAmelCase = sampling_rate _UpperCAmelCase = audio_channels _UpperCAmelCase = normalize _UpperCAmelCase = chunk_length_s _UpperCAmelCase = overlap _UpperCAmelCase = hidden_size _UpperCAmelCase = num_filters _UpperCAmelCase = num_residual_layers _UpperCAmelCase = upsampling_ratios _UpperCAmelCase = norm_type _UpperCAmelCase = kernel_size _UpperCAmelCase = last_kernel_size _UpperCAmelCase = residual_kernel_size _UpperCAmelCase = dilation_growth_rate _UpperCAmelCase = use_causal_conv _UpperCAmelCase = pad_mode _UpperCAmelCase = compress _UpperCAmelCase = num_lstm_layers _UpperCAmelCase = trim_right_ratio _UpperCAmelCase = codebook_size _UpperCAmelCase = codebook_dim if codebook_dim is not None else hidden_size _UpperCAmelCase = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( f'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**snake_case_ ) @property def lowercase ( self : Optional[int] ): if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def lowercase ( self : Tuple ): if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def lowercase ( self : Dict ): _UpperCAmelCase = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def lowercase ( self : Union[str, Any] ): return int(1_0_0_0 * self.target_bandwidths[-1] // (self.frame_rate * 1_0) )
22
'''simple docstring''' import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( "kwargs, expected" , [ ({"num_shards": 0, "max_num_jobs": 1}, []), ({"num_shards": 10, "max_num_jobs": 1}, [range(10 )]), ({"num_shards": 10, "max_num_jobs": 10}, [range(__lowercase , i + 1 ) for i in range(10 )]), ({"num_shards": 1, "max_num_jobs": 10}, [range(1 )]), ({"num_shards": 10, "max_num_jobs": 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({"num_shards": 3, "max_num_jobs": 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def UpperCAmelCase_ ( __lowercase : int , __lowercase : Dict ) -> List[Any]: '''simple docstring''' _UpperCAmelCase = _distribute_shards(**__lowercase ) assert out == expected @pytest.mark.parametrize( "gen_kwargs, max_num_jobs, expected" , [ ({"foo": 0}, 10, [{"foo": 0}]), ({"shards": [0, 1, 2, 3]}, 1, [{"shards": [0, 1, 2, 3]}]), ({"shards": [0, 1, 2, 3]}, 4, [{"shards": [0]}, {"shards": [1]}, {"shards": [2]}, {"shards": [3]}]), ({"shards": [0, 1]}, 4, [{"shards": [0]}, {"shards": [1]}]), ({"shards": [0, 1, 2, 3]}, 2, [{"shards": [0, 1]}, {"shards": [2, 3]}]), ] , ) def UpperCAmelCase_ ( __lowercase : Dict , __lowercase : Optional[Any] , __lowercase : int ) -> str: '''simple docstring''' _UpperCAmelCase = _split_gen_kwargs(__lowercase , __lowercase ) assert out == expected @pytest.mark.parametrize( "gen_kwargs, expected" , [ ({"foo": 0}, 1), ({"shards": [0]}, 1), ({"shards": [0, 1, 2, 3]}, 4), ({"shards": [0, 1, 2, 3], "foo": 0}, 4), ({"shards": [0, 1, 2, 3], "other": (0, 1)}, 4), ({"shards": [0, 1, 2, 3], "shards2": [0, 1]}, RuntimeError), ] , ) def UpperCAmelCase_ ( __lowercase : Optional[Any] , __lowercase : List[Any] ) -> List[Any]: '''simple docstring''' if expected is RuntimeError: with pytest.raises(__lowercase ): _number_of_shards_in_gen_kwargs(__lowercase ) else: _UpperCAmelCase = _number_of_shards_in_gen_kwargs(__lowercase ) assert out == expected
22
1
"""simple docstring""" import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument("""--user""", type=str, default="""ubuntu""") parser.add_argument("""--host""", type=str, default="""localhost""") parser.add_argument("""--key_path""", type=str, default=None) parser.add_argument("""--instance""", type=str, default="""V100:1""") parser.add_argument("""--provider""", type=str, default="""cheapest""") parser.add_argument("""--use_spot""", type=bool, default=False) parser.add_argument("""--example""", type=str, default="""pytorch/text-generation/run_generation.py""") lowerCamelCase__ , lowerCamelCase__ = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("""Cannot specify both BYO and on-demand cluster args""") lowerCamelCase__ = rh.cluster( name="""rh-cluster""", ips=[args.host], ssh_creds={"""ssh_user""": args.user, """ssh_private_key""": args.key_path} ) else: lowerCamelCase__ = rh.cluster( name="""rh-cluster""", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) lowerCamelCase__ = args.example.rsplit("""/""", 1)[0] # Set up remote environment cluster.install_packages(["""pip:./"""]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([f'pip install -r transformers/examples/{example_dir}/requirements.txt']) cluster.run(["""pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"""]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([f'python transformers/examples/{args.example} {" ".join(shlex.quote(arg) for arg in unknown)}']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
182
"""simple docstring""" import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize('dataset_size' , [None, 400 * 2**20, 600 * 2**20] ) @pytest.mark.parametrize('input_in_memory_max_size' , ['default', 0, 100 * 2**20, 900 * 2**20] ) def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): if input_in_memory_max_size != "default": monkeypatch.setattr(datasets.config , 'IN_MEMORY_MAX_SIZE' , _UpperCamelCase ) __lowerCAmelCase : Union[str, Any] = datasets.config.IN_MEMORY_MAX_SIZE if input_in_memory_max_size == "default": assert in_memory_max_size == 0 else: assert in_memory_max_size == input_in_memory_max_size if dataset_size and in_memory_max_size: __lowerCAmelCase : Tuple = dataset_size < in_memory_max_size else: __lowerCAmelCase : str = False __lowerCAmelCase : Optional[int] = is_small_dataset(_UpperCamelCase ) assert result == expected
182
1
import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) from transformers.utils import logging logging.set_verbosity_info() __a = logging.get_logger(__name__) def a ( snake_case__: Optional[Any] ): '''simple docstring''' lowercase_ = MobileNetVaConfig(layer_norm_eps=0.0_0_1 ) if "_quant" in model_name: raise ValueError('''Quantized models are not supported.''' ) lowercase_ = re.match(r'''^mobilenet_v1_([^_]*)_([^_]*)$''' , snake_case__ ) if matches: lowercase_ = float(matches[1] ) lowercase_ = int(matches[2] ) # The TensorFlow version of MobileNetV1 predicts 1001 classes instead of # the usual 1000. The first class (index 0) is "background". lowercase_ = 1_001 lowercase_ = '''imagenet-1k-id2label.json''' lowercase_ = '''huggingface/label-files''' lowercase_ = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type='''dataset''' ) , '''r''' ) ) lowercase_ = {int(snake_case__ ) + 1: v for k, v in idalabel.items()} lowercase_ = '''background''' lowercase_ = idalabel lowercase_ = {v: k for k, v in idalabel.items()} return config def a ( ): '''simple docstring''' lowercase_ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowercase_ = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) return im @torch.no_grad() def a ( snake_case__: Optional[int] , snake_case__: int , snake_case__: Dict , snake_case__: Optional[Any]=False ): '''simple docstring''' lowercase_ = get_mobilenet_va_config(snake_case__ ) # Load 🤗 model lowercase_ = MobileNetVaForImageClassification(snake_case__ ).eval() # Load weights from TensorFlow checkpoint load_tf_weights_in_mobilenet_va(snake_case__ , snake_case__ , snake_case__ ) # Check outputs on an image, prepared by MobileNetV1ImageProcessor lowercase_ = MobileNetVaImageProcessor( crop_size={'''width''': config.image_size, '''height''': config.image_size} , size={'''shortest_edge''': config.image_size + 32} , ) lowercase_ = image_processor(images=prepare_img() , return_tensors='''pt''' ) lowercase_ = model(**snake_case__ ) lowercase_ = outputs.logits assert logits.shape == (1, 1_001) if model_name == "mobilenet_v1_1.0_224": lowercase_ = torch.tensor([-4.1_7_3_9, -1.1_2_3_3, 3.1_2_0_5] ) elif model_name == "mobilenet_v1_0.75_192": lowercase_ = torch.tensor([-3.9_4_4_0, -2.3_1_4_1, -0.3_3_3_3] ) else: lowercase_ = None if expected_logits is not None: assert torch.allclose(logits[0, :3] , snake_case__ , atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: print('''Pushing to the hub...''' ) lowercase_ = '''google/''' + model_name image_processor.push_to_hub(snake_case__ ) model.push_to_hub(snake_case__ ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='mobilenet_v1_1.0_224', type=str, help='Name of the MobileNetV1 model you\'d like to convert. Should in the form \'mobilenet_v1_<depth>_<size>\'.', ) parser.add_argument( '--checkpoint_path', required=True, type=str, help='Path to the original TensorFlow checkpoint (.ckpt file).' ) parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) __a = parser.parse_args() convert_movilevit_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
30
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available _lowerCamelCase ={ "configuration_ernie": ["ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ErnieConfig", "ErnieOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase =[ "ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST", "ErnieForCausalLM", "ErnieForMaskedLM", "ErnieForMultipleChoice", "ErnieForNextSentencePrediction", "ErnieForPreTraining", "ErnieForQuestionAnswering", "ErnieForSequenceClassification", "ErnieForTokenClassification", "ErnieModel", "ErniePreTrainedModel", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys _lowerCamelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
334
0
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __magic_name__ = logging.get_logger(__name__) __magic_name__ = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def _lowerCAmelCase ( A__: Any , A__: str , A__: Tuple , A__: Optional[Any] , A__: List[Any] ): '''simple docstring''' for attribute in key.split('''.''' ): UpperCAmelCase = getattr(__lowerCAmelCase , __lowerCAmelCase ) if weight_type is not None: UpperCAmelCase = getattr(__lowerCAmelCase , __lowerCAmelCase ).shape else: UpperCAmelCase = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase = value elif weight_type == "weight_g": UpperCAmelCase = value elif weight_type == "weight_v": UpperCAmelCase = value elif weight_type == "bias": UpperCAmelCase = value else: UpperCAmelCase = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def _lowerCAmelCase ( A__: int , A__: Dict , A__: List[str] ): '''simple docstring''' UpperCAmelCase = [] UpperCAmelCase = fairseq_model.state_dict() UpperCAmelCase = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase = False if "conv_layers" in name: load_conv_layer( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) UpperCAmelCase = True else: for key, mapped_key in MAPPING.items(): UpperCAmelCase = '''sew.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: UpperCAmelCase = True if "*" in mapped_key: UpperCAmelCase = name.split(__lowerCAmelCase )[0].split('''.''' )[-2] UpperCAmelCase = mapped_key.replace('''*''' , __lowerCAmelCase ) if "weight_g" in name: UpperCAmelCase = '''weight_g''' elif "weight_v" in name: UpperCAmelCase = '''weight_v''' elif "weight" in name: UpperCAmelCase = '''weight''' elif "bias" in name: UpperCAmelCase = '''bias''' else: UpperCAmelCase = None set_recursively(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) continue if not is_used: unused_weights.append(__lowerCAmelCase ) logger.warning(F"""Unused weights: {unused_weights}""" ) def _lowerCAmelCase ( A__: Optional[Any] , A__: str , A__: Any , A__: Optional[int] , A__: int ): '''simple docstring''' UpperCAmelCase = full_name.split('''conv_layers.''' )[-1] UpperCAmelCase = name.split('''.''' ) UpperCAmelCase = int(items[0] ) UpperCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) UpperCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__lowerCAmelCase ) def _lowerCAmelCase ( A__: Tuple , A__: Dict ): '''simple docstring''' UpperCAmelCase = SEWConfig() if is_finetuned: UpperCAmelCase = model.wav_encoder.wav_model.cfg else: UpperCAmelCase = model.cfg UpperCAmelCase = fs_config.conv_bias UpperCAmelCase = eval(fs_config.conv_feature_layers ) UpperCAmelCase = [x[0] for x in conv_layers] UpperCAmelCase = [x[1] for x in conv_layers] UpperCAmelCase = [x[2] for x in conv_layers] UpperCAmelCase = '''gelu''' UpperCAmelCase = '''layer''' if fs_config.extractor_mode == '''layer_norm''' else '''group''' UpperCAmelCase = 0.0 UpperCAmelCase = fs_config.activation_fn.name UpperCAmelCase = fs_config.encoder_embed_dim UpperCAmelCase = 0.02 UpperCAmelCase = fs_config.encoder_ffn_embed_dim UpperCAmelCase = 1E-5 UpperCAmelCase = fs_config.encoder_layerdrop UpperCAmelCase = fs_config.encoder_attention_heads UpperCAmelCase = fs_config.conv_pos_groups UpperCAmelCase = fs_config.conv_pos UpperCAmelCase = len(__lowerCAmelCase ) UpperCAmelCase = fs_config.encoder_layers UpperCAmelCase = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: UpperCAmelCase = model.cfg UpperCAmelCase = fs_config.final_dropout UpperCAmelCase = fs_config.layerdrop UpperCAmelCase = fs_config.activation_dropout UpperCAmelCase = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 UpperCAmelCase = fs_config.attention_dropout UpperCAmelCase = fs_config.dropout_input UpperCAmelCase = fs_config.dropout UpperCAmelCase = fs_config.mask_channel_length UpperCAmelCase = fs_config.mask_channel_prob UpperCAmelCase = fs_config.mask_length UpperCAmelCase = fs_config.mask_prob UpperCAmelCase = '''Wav2Vec2FeatureExtractor''' UpperCAmelCase = '''Wav2Vec2CTCTokenizer''' return config @torch.no_grad() def _lowerCAmelCase ( A__: List[str] , A__: List[str] , A__: Optional[int]=None , A__: Union[str, Any]=None , A__: Dict=True ): '''simple docstring''' if is_finetuned: UpperCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: UpperCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: UpperCAmelCase = SEWConfig.from_pretrained(__lowerCAmelCase ) else: UpperCAmelCase = convert_config(model[0] , __lowerCAmelCase ) UpperCAmelCase = model[0].eval() UpperCAmelCase = True if config.feat_extract_norm == '''layer''' else False UpperCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=__lowerCAmelCase , return_attention_mask=__lowerCAmelCase , ) if is_finetuned: if dict_path: UpperCAmelCase = Dictionary.load(__lowerCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase = target_dict.pad_index UpperCAmelCase = target_dict.bos_index UpperCAmelCase = target_dict.pad_index UpperCAmelCase = target_dict.bos_index UpperCAmelCase = target_dict.eos_index UpperCAmelCase = len(target_dict.symbols ) UpperCAmelCase = os.path.join(__lowerCAmelCase , '''vocab.json''' ) if not os.path.isdir(__lowerCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__lowerCAmelCase ) ) return os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , __lowerCAmelCase ) UpperCAmelCase = WavaVecaCTCTokenizer( __lowerCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=__lowerCAmelCase , ) UpperCAmelCase = WavaVecaProcessor(feature_extractor=__lowerCAmelCase , tokenizer=__lowerCAmelCase ) processor.save_pretrained(__lowerCAmelCase ) UpperCAmelCase = SEWForCTC(__lowerCAmelCase ) else: UpperCAmelCase = SEWModel(__lowerCAmelCase ) feature_extractor.save_pretrained(__lowerCAmelCase ) recursively_load_weights(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) hf_model.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": __magic_name__ = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) __magic_name__ = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
362
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging __magic_name__ = logging.get_logger(__name__) def _lowerCAmelCase ( A__: nn.ModuleList , A__: nn.ModuleList , A__: List[int] ): '''simple docstring''' UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(A__ ) == len(A__ ), F"""{len(A__ )} != {len(A__ )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) __magic_name__ = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } __magic_name__ = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _lowerCAmelCase ( A__: List[str] , A__: Optional[int] ): '''simple docstring''' try: UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(A__ ) ) def _lowerCAmelCase ( A__: Optional[int] , A__: Tuple ): '''simple docstring''' if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(A__ ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _lowerCAmelCase ( A__: Union[str, PreTrainedModel] , A__: Union[str, Path] = "student" , A__: Union[int, None] = None , A__: Union[int, None] = None , A__: Optional[int]=False , A__: Tuple=None , A__: Any=None , **A__: List[str] , ): '''simple docstring''' UpperCAmelCase = '''encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.''' assert (e is not None) or (d is not None), _msg if isinstance(A__ , A__ ): AutoTokenizer.from_pretrained(A__ ).save_pretrained(A__ ) # purely for convenience UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(A__ ).eval() else: assert isinstance(A__ , A__ ), F"""teacher must be a model or string got type {type(A__ )}""" UpperCAmelCase = teacher.config.to_diff_dict() try: UpperCAmelCase , UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: UpperCAmelCase = teacher_e if d is None: UpperCAmelCase = teacher_d init_kwargs.update({'''encoder_layers''': e, '''decoder_layers''': d} ) except AttributeError: # T5 if hasattr(teacher.config , '''num_encoder_layers''' ): UpperCAmelCase , UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: UpperCAmelCase , UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: UpperCAmelCase = teacher_e if d is None: UpperCAmelCase = teacher_d if hasattr(teacher.config , '''num_encoder_layers''' ): init_kwargs.update({'''num_encoder_layers''': e, '''num_decoder_layers''': d} ) else: init_kwargs.update({'''num_layers''': e, '''num_decoder_layers''': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(A__ ) # Copy weights UpperCAmelCase = teacher.config_class(**A__ ) UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(A__ ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=A__ ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save UpperCAmelCase , UpperCAmelCase = list(range(A__ ) ), list(range(A__ ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(A__ ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: UpperCAmelCase = pick_layers_to_copy(A__ , A__ ) if d_layers_to_copy is None: UpperCAmelCase = pick_layers_to_copy(A__ , A__ ) try: if hasattr( A__ , '''prophetnet''' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , A__ ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , A__ ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , A__ ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , A__ ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , A__ ) copy_layers(teacher.decoder.block , student.decoder.block , A__ ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) UpperCAmelCase = { '''teacher_type''': teacher.config.model_type, '''copied_encoder_layers''': e_layers_to_copy, '''copied_decoder_layers''': d_layers_to_copy, } student.save_pretrained(A__ ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
152
0
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig A__ : str = logging.get_logger(__name__) # General docstring A__ : List[str] = '''ResNetConfig''' # Base docstring A__ : Tuple = '''microsoft/resnet-50''' A__ : List[str] = [1, 2048, 7, 7] # Image classification docstring A__ : str = '''microsoft/resnet-50''' A__ : List[str] = '''tiger cat''' A__ : Dict = [ '''microsoft/resnet-50''', # See all resnet models at https://huggingface.co/models?filter=resnet ] class __snake_case ( nn.Module ): def __init__( self : int , A_ : int , A_ : int , A_ : int = 3 , A_ : int = 1 , A_ : str = "relu"): super().__init__() lowerCAmelCase_ : Optional[Any] = nn.Convad( A_ , A_ , kernel_size=A_ , stride=A_ , padding=kernel_size // 2 , bias=A_) lowerCAmelCase_ : Any = nn.BatchNormad(A_) lowerCAmelCase_ : Optional[int] = ACTaFN[activation] if activation is not None else nn.Identity() def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Tensor): lowerCAmelCase_ : Tuple = self.convolution(A_) lowerCAmelCase_ : str = self.normalization(A_) lowerCAmelCase_ : Tuple = self.activation(A_) return hidden_state class __snake_case ( nn.Module ): def __init__( self : Optional[Any] , A_ : ResNetConfig): super().__init__() lowerCAmelCase_ : Union[str, Any] = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act) lowerCAmelCase_ : Any = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1) lowerCAmelCase_ : Dict = config.num_channels def UpperCAmelCase__ ( self : Any , A_ : Tensor): lowerCAmelCase_ : Tuple = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''') lowerCAmelCase_ : int = self.embedder(A_) lowerCAmelCase_ : List[str] = self.pooler(A_) return embedding class __snake_case ( nn.Module ): def __init__( self : List[str] , A_ : int , A_ : int , A_ : int = 2): super().__init__() lowerCAmelCase_ : Optional[Any] = nn.Convad(A_ , A_ , kernel_size=1 , stride=A_ , bias=A_) lowerCAmelCase_ : int = nn.BatchNormad(A_) def UpperCAmelCase__ ( self : Any , A_ : Tensor): lowerCAmelCase_ : str = self.convolution(A_) lowerCAmelCase_ : Optional[int] = self.normalization(A_) return hidden_state class __snake_case ( nn.Module ): def __init__( self : int , A_ : int , A_ : int , A_ : int = 1 , A_ : str = "relu"): super().__init__() lowerCAmelCase_ : Tuple = in_channels != out_channels or stride != 1 lowerCAmelCase_ : str = ( ResNetShortCut(A_ , A_ , stride=A_) if should_apply_shortcut else nn.Identity() ) lowerCAmelCase_ : str = nn.Sequential( ResNetConvLayer(A_ , A_ , stride=A_) , ResNetConvLayer(A_ , A_ , activation=A_) , ) lowerCAmelCase_ : List[Any] = ACTaFN[activation] def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Dict): lowerCAmelCase_ : Optional[int] = hidden_state lowerCAmelCase_ : Any = self.layer(A_) lowerCAmelCase_ : Optional[int] = self.shortcut(A_) hidden_state += residual lowerCAmelCase_ : Tuple = self.activation(A_) return hidden_state class __snake_case ( nn.Module ): def __init__( self : Optional[int] , A_ : int , A_ : int , A_ : int = 1 , A_ : str = "relu" , A_ : int = 4): super().__init__() lowerCAmelCase_ : Union[str, Any] = in_channels != out_channels or stride != 1 lowerCAmelCase_ : List[str] = out_channels // reduction lowerCAmelCase_ : Union[str, Any] = ( ResNetShortCut(A_ , A_ , stride=A_) if should_apply_shortcut else nn.Identity() ) lowerCAmelCase_ : List[Any] = nn.Sequential( ResNetConvLayer(A_ , A_ , kernel_size=1) , ResNetConvLayer(A_ , A_ , stride=A_) , ResNetConvLayer(A_ , A_ , kernel_size=1 , activation=A_) , ) lowerCAmelCase_ : int = ACTaFN[activation] def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Dict): lowerCAmelCase_ : List[str] = hidden_state lowerCAmelCase_ : Optional[Any] = self.layer(A_) lowerCAmelCase_ : Optional[Any] = self.shortcut(A_) hidden_state += residual lowerCAmelCase_ : Dict = self.activation(A_) return hidden_state class __snake_case ( nn.Module ): def __init__( self : int , A_ : ResNetConfig , A_ : int , A_ : int , A_ : int = 2 , A_ : int = 2 , ): super().__init__() lowerCAmelCase_ : List[str] = ResNetBottleNeckLayer if config.layer_type == '''bottleneck''' else ResNetBasicLayer lowerCAmelCase_ : int = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(A_ , A_ , stride=A_ , activation=config.hidden_act) , *[layer(A_ , A_ , activation=config.hidden_act) for _ in range(depth - 1)] , ) def UpperCAmelCase__ ( self : Tuple , A_ : Tensor): lowerCAmelCase_ : Union[str, Any] = input for layer in self.layers: lowerCAmelCase_ : List[str] = layer(A_) return hidden_state class __snake_case ( nn.Module ): def __init__( self : List[str] , A_ : ResNetConfig): super().__init__() lowerCAmelCase_ : Union[str, Any] = nn.ModuleList([]) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( A_ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , )) lowerCAmelCase_ : Tuple = zip(config.hidden_sizes , config.hidden_sizes[1:]) for (in_channels, out_channels), depth in zip(A_ , config.depths[1:]): self.stages.append(ResNetStage(A_ , A_ , A_ , depth=A_)) def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Tensor , A_ : bool = False , A_ : bool = True): lowerCAmelCase_ : Union[str, Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowerCAmelCase_ : int = hidden_states + (hidden_state,) lowerCAmelCase_ : int = stage_module(A_) if output_hidden_states: lowerCAmelCase_ : List[Any] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None) return BaseModelOutputWithNoAttention( last_hidden_state=A_ , hidden_states=A_ , ) class __snake_case ( UpperCamelCase_ ): _a = ResNetConfig _a = '''resnet''' _a = '''pixel_values''' _a = True def UpperCAmelCase__ ( self : Optional[Any] , A_ : Any): if isinstance(A_ , nn.Convad): nn.init.kaiming_normal_(module.weight , mode='''fan_out''' , nonlinearity='''relu''') elif isinstance(A_ , (nn.BatchNormad, nn.GroupNorm)): nn.init.constant_(module.weight , 1) nn.init.constant_(module.bias , 0) def UpperCAmelCase__ ( self : str , A_ : int , A_ : Union[str, Any]=False): if isinstance(A_ , A_): lowerCAmelCase_ : Union[str, Any] = value A__ : Any = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`ResNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' A__ : Union[str, Any] = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( '''The bare ResNet model outputting raw features without any specific head on top.''' ,UpperCamelCase_ ,) class __snake_case ( UpperCamelCase_ ): def __init__( self : Any , A_ : Any): super().__init__(A_) lowerCAmelCase_ : str = config lowerCAmelCase_ : Union[str, Any] = ResNetEmbeddings(A_) lowerCAmelCase_ : Dict = ResNetEncoder(A_) lowerCAmelCase_ : Any = nn.AdaptiveAvgPoolad((1, 1)) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=A_ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCAmelCase__ ( self : Optional[int] , A_ : Tensor , A_ : Optional[bool] = None , A_ : Optional[bool] = None): lowerCAmelCase_ : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase_ : Optional[int] = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase_ : Tuple = self.embedder(A_) lowerCAmelCase_ : Tuple = self.encoder( A_ , output_hidden_states=A_ , return_dict=A_) lowerCAmelCase_ : List[Any] = encoder_outputs[0] lowerCAmelCase_ : str = self.pooler(A_) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=A_ , pooler_output=A_ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( ''' ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' ,UpperCamelCase_ ,) class __snake_case ( UpperCamelCase_ ): def __init__( self : Dict , A_ : List[str]): super().__init__(A_) lowerCAmelCase_ : Optional[Any] = config.num_labels lowerCAmelCase_ : Union[str, Any] = ResNetModel(A_) # classification head lowerCAmelCase_ : List[str] = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=A_ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCAmelCase__ ( self : int , A_ : Optional[torch.FloatTensor] = None , A_ : Optional[torch.LongTensor] = None , A_ : Optional[bool] = None , A_ : Optional[bool] = None , ): lowerCAmelCase_ : Dict = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase_ : Dict = self.resnet(A_ , output_hidden_states=A_ , return_dict=A_) lowerCAmelCase_ : Optional[Any] = outputs.pooler_output if return_dict else outputs[1] lowerCAmelCase_ : List[str] = self.classifier(A_) lowerCAmelCase_ : List[str] = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowerCAmelCase_ : List[Any] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowerCAmelCase_ : Optional[Any] = '''single_label_classification''' else: lowerCAmelCase_ : str = '''multi_label_classification''' if self.config.problem_type == "regression": lowerCAmelCase_ : List[str] = MSELoss() if self.num_labels == 1: lowerCAmelCase_ : List[Any] = loss_fct(logits.squeeze() , labels.squeeze()) else: lowerCAmelCase_ : Dict = loss_fct(A_ , A_) elif self.config.problem_type == "single_label_classification": lowerCAmelCase_ : int = CrossEntropyLoss() lowerCAmelCase_ : Optional[int] = loss_fct(logits.view(-1 , self.num_labels) , labels.view(-1)) elif self.config.problem_type == "multi_label_classification": lowerCAmelCase_ : List[str] = BCEWithLogitsLoss() lowerCAmelCase_ : Tuple = loss_fct(A_ , A_) if not return_dict: lowerCAmelCase_ : int = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=A_ , logits=A_ , hidden_states=outputs.hidden_states) @add_start_docstrings( ''' ResNet backbone, to be used with frameworks like DETR and MaskFormer. ''' ,UpperCamelCase_ ,) class __snake_case ( UpperCamelCase_ ,UpperCamelCase_ ): def __init__( self : Dict , A_ : int): super().__init__(A_) super()._init_backbone(A_) lowerCAmelCase_ : List[Any] = [config.embedding_size] + config.hidden_sizes lowerCAmelCase_ : Dict = ResNetEmbeddings(A_) lowerCAmelCase_ : Any = ResNetEncoder(A_) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(A_) @replace_return_docstrings(output_type=A_ , config_class=_CONFIG_FOR_DOC) def UpperCAmelCase__ ( self : List[str] , A_ : Tensor , A_ : Optional[bool] = None , A_ : Optional[bool] = None): lowerCAmelCase_ : Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase_ : Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase_ : Any = self.embedder(A_) lowerCAmelCase_ : str = self.encoder(A_ , output_hidden_states=A_ , return_dict=A_) lowerCAmelCase_ : Union[str, Any] = outputs.hidden_states lowerCAmelCase_ : List[Any] = () for idx, stage in enumerate(self.stage_names): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: lowerCAmelCase_ : Any = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=A_ , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=A_ , )
103
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() lowerCAmelCase : Any = logging.get_logger(__name__) def A_ ( a , a , a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = WavaVecaForSequenceClassification.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE_ : Optional[Any] = downstream_dict['projector.weight'] SCREAMING_SNAKE_CASE_ : Union[str, Any] = downstream_dict['projector.bias'] SCREAMING_SNAKE_CASE_ : List[str] = downstream_dict['model.post_net.linear.weight'] SCREAMING_SNAKE_CASE_ : Optional[Any] = downstream_dict['model.post_net.linear.bias'] return model def A_ ( a , a , a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = WavaVecaForAudioFrameClassification.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE_ : Dict = downstream_dict['model.linear.weight'] SCREAMING_SNAKE_CASE_ : List[Any] = downstream_dict['model.linear.bias'] return model def A_ ( a , a , a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = WavaVecaForXVector.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE_ : Tuple = downstream_dict['connector.weight'] SCREAMING_SNAKE_CASE_ : Any = downstream_dict['connector.bias'] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): SCREAMING_SNAKE_CASE_ : Optional[Any] = downstream_dict[ f"model.framelevel_feature_extractor.module.{i}.kernel.weight" ] SCREAMING_SNAKE_CASE_ : Optional[Any] = downstream_dict[f"model.framelevel_feature_extractor.module.{i}.kernel.bias"] SCREAMING_SNAKE_CASE_ : Optional[Any] = downstream_dict['model.utterancelevel_feature_extractor.linear1.weight'] SCREAMING_SNAKE_CASE_ : Tuple = downstream_dict['model.utterancelevel_feature_extractor.linear1.bias'] SCREAMING_SNAKE_CASE_ : Any = downstream_dict['model.utterancelevel_feature_extractor.linear2.weight'] SCREAMING_SNAKE_CASE_ : List[str] = downstream_dict['model.utterancelevel_feature_extractor.linear2.bias'] SCREAMING_SNAKE_CASE_ : Optional[int] = downstream_dict['objective.W'] return model @torch.no_grad() def A_ ( a , a , a , a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = torch.load(a , map_location='cpu' ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = checkpoint['Downstream'] SCREAMING_SNAKE_CASE_ : Optional[int] = WavaVecaConfig.from_pretrained(a ) SCREAMING_SNAKE_CASE_ : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained( a , return_attention_mask=a , do_normalize=a ) SCREAMING_SNAKE_CASE_ : Tuple = hf_config.architectures[0] if arch.endswith('ForSequenceClassification' ): SCREAMING_SNAKE_CASE_ : Tuple = convert_classification(a , a , a ) elif arch.endswith('ForAudioFrameClassification' ): SCREAMING_SNAKE_CASE_ : str = convert_diarization(a , a , a ) elif arch.endswith('ForXVector' ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = convert_xvector(a , a , a ) else: raise NotImplementedError(f"S3PRL weights conversion is not supported for {arch}" ) if hf_config.use_weighted_layer_sum: SCREAMING_SNAKE_CASE_ : Dict = checkpoint['Featurizer']['weights'] hf_feature_extractor.save_pretrained(a ) hf_model.save_pretrained(a ) if __name__ == "__main__": lowerCAmelCase : List[Any] = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') lowerCAmelCase : List[str] = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
253
0
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() _a : str= logging.get_logger(__name__) _a : List[Any]= { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def __UpperCAmelCase ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Any ) -> Union[str, Any]: '''simple docstring''' for attribute in key.split('.' ): __snake_case : List[Any] = getattr(UpperCAmelCase_ , UpperCAmelCase_ ) if weight_type is not None: __snake_case : Optional[int] = getattr(UpperCAmelCase_ , UpperCAmelCase_ ).shape else: __snake_case : List[Any] = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": __snake_case : Union[str, Any] = value elif weight_type == "weight_g": __snake_case : List[Any] = value elif weight_type == "weight_v": __snake_case : str = value elif weight_type == "bias": __snake_case : Union[str, Any] = value else: __snake_case : Optional[int] = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def __UpperCAmelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' __snake_case : Tuple = [] __snake_case : Any = fairseq_model.state_dict() __snake_case : List[Any] = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __snake_case : Dict = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , hf_model.config.feat_extract_norm == 'group' , ) __snake_case : Optional[int] = True else: for key, mapped_key in MAPPING.items(): __snake_case : List[str] = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: __snake_case : str = True if "*" in mapped_key: __snake_case : Optional[int] = name.split(UpperCAmelCase_ )[0].split('.' )[-2] __snake_case : Optional[int] = mapped_key.replace('*' , UpperCAmelCase_ ) if "weight_g" in name: __snake_case : int = 'weight_g' elif "weight_v" in name: __snake_case : Any = 'weight_v' elif "weight" in name: __snake_case : Union[str, Any] = 'weight' elif "bias" in name: __snake_case : List[Any] = 'bias' else: __snake_case : List[str] = None set_recursively(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) continue if not is_used: unused_weights.append(UpperCAmelCase_ ) logger.warning(F"Unused weights: {unused_weights}" ) def __UpperCAmelCase ( UpperCAmelCase_ : str , UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Dict ) -> str: '''simple docstring''' __snake_case : str = full_name.split('conv_layers.' )[-1] __snake_case : int = name.split('.' ) __snake_case : int = int(items[0] ) __snake_case : int = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) __snake_case : List[Any] = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) __snake_case : int = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) __snake_case : Optional[int] = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) __snake_case : Any = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(UpperCAmelCase_ ) def __UpperCAmelCase ( UpperCAmelCase_ : Any , UpperCAmelCase_ : Dict ) -> Any: '''simple docstring''' __snake_case : Union[str, Any] = SEWConfig() if is_finetuned: __snake_case : int = model.wav_encoder.wav_model.cfg else: __snake_case : List[Any] = model.cfg __snake_case : Optional[Any] = fs_config.conv_bias __snake_case : Any = eval(fs_config.conv_feature_layers ) __snake_case : str = [x[0] for x in conv_layers] __snake_case : Dict = [x[1] for x in conv_layers] __snake_case : List[Any] = [x[2] for x in conv_layers] __snake_case : Any = 'gelu' __snake_case : int = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' __snake_case : Any = 0.0 __snake_case : Optional[Any] = fs_config.activation_fn.name __snake_case : Any = fs_config.encoder_embed_dim __snake_case : Tuple = 0.02 __snake_case : int = fs_config.encoder_ffn_embed_dim __snake_case : str = 1E-5 __snake_case : Any = fs_config.encoder_layerdrop __snake_case : int = fs_config.encoder_attention_heads __snake_case : Optional[Any] = fs_config.conv_pos_groups __snake_case : List[str] = fs_config.conv_pos __snake_case : List[str] = len(UpperCAmelCase_ ) __snake_case : int = fs_config.encoder_layers __snake_case : int = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: __snake_case : int = model.cfg __snake_case : List[Any] = fs_config.final_dropout __snake_case : Optional[Any] = fs_config.layerdrop __snake_case : Optional[Any] = fs_config.activation_dropout __snake_case : List[str] = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 __snake_case : int = fs_config.attention_dropout __snake_case : Tuple = fs_config.dropout_input __snake_case : Union[str, Any] = fs_config.dropout __snake_case : List[Any] = fs_config.mask_channel_length __snake_case : Dict = fs_config.mask_channel_prob __snake_case : Tuple = fs_config.mask_length __snake_case : Tuple = fs_config.mask_prob __snake_case : Tuple = 'Wav2Vec2FeatureExtractor' __snake_case : List[Any] = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def __UpperCAmelCase ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[Any]=None , UpperCAmelCase_ : Tuple=None , UpperCAmelCase_ : Dict=True ) -> Union[str, Any]: '''simple docstring''' if is_finetuned: __snake_case , __snake_case , __snake_case : List[Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: __snake_case , __snake_case , __snake_case : List[Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: __snake_case : List[str] = SEWConfig.from_pretrained(UpperCAmelCase_ ) else: __snake_case : Optional[Any] = convert_config(model[0] , UpperCAmelCase_ ) __snake_case : Dict = model[0].eval() __snake_case : List[str] = True if config.feat_extract_norm == 'layer' else False __snake_case : Tuple = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=UpperCAmelCase_ , return_attention_mask=UpperCAmelCase_ , ) if is_finetuned: if dict_path: __snake_case : Optional[Any] = Dictionary.load(UpperCAmelCase_ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __snake_case : Tuple = target_dict.pad_index __snake_case : Tuple = target_dict.bos_index __snake_case : List[Any] = target_dict.pad_index __snake_case : List[Any] = target_dict.bos_index __snake_case : Tuple = target_dict.eos_index __snake_case : Dict = len(target_dict.symbols ) __snake_case : List[Any] = os.path.join(UpperCAmelCase_ , 'vocab.json' ) if not os.path.isdir(UpperCAmelCase_ ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(UpperCAmelCase_ ) ) return os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) with open(UpperCAmelCase_ , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , UpperCAmelCase_ ) __snake_case : Tuple = WavaVecaCTCTokenizer( UpperCAmelCase_ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=UpperCAmelCase_ , ) __snake_case : List[Any] = WavaVecaProcessor(feature_extractor=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) processor.save_pretrained(UpperCAmelCase_ ) __snake_case : int = SEWForCTC(UpperCAmelCase_ ) else: __snake_case : str = SEWModel(UpperCAmelCase_ ) feature_extractor.save_pretrained(UpperCAmelCase_ ) recursively_load_weights(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) hf_model.save_pretrained(UpperCAmelCase_ ) if __name__ == "__main__": _a : Any= argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) _a : List[Any]= parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
95
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer _a : List[Any]= logging.get_logger(__name__) _a : Any= {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _a : int= { "vocab_file": { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt" ), } } _a : Optional[Any]= { "junnyu/roformer_chinese_small": 1_536, "junnyu/roformer_chinese_base": 1_536, "junnyu/roformer_chinese_char_small": 512, "junnyu/roformer_chinese_char_base": 512, "junnyu/roformer_small_discriminator": 128, "junnyu/roformer_small_generator": 128, } _a : str= { "junnyu/roformer_chinese_small": {"do_lower_case": True}, "junnyu/roformer_chinese_base": {"do_lower_case": True}, "junnyu/roformer_chinese_char_small": {"do_lower_case": True}, "junnyu/roformer_chinese_char_base": {"do_lower_case": True}, "junnyu/roformer_small_discriminator": {"do_lower_case": True}, "junnyu/roformer_small_generator": {"do_lower_case": True}, } class UpperCamelCase ( lowercase ): UpperCAmelCase : Union[str, Any] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : int = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase : Dict = RoFormerTokenizer def __init__(self : List[Any] , _A : Any=None , _A : int=None , _A : Dict=True , _A : List[Any]="[UNK]" , _A : Tuple="[SEP]" , _A : List[Any]="[PAD]" , _A : str="[CLS]" , _A : int="[MASK]" , _A : Optional[int]=True , _A : List[str]=None , **_A : int , ) -> Dict: super().__init__( _A , tokenizer_file=_A , do_lower_case=_A , unk_token=_A , sep_token=_A , pad_token=_A , cls_token=_A , mask_token=_A , tokenize_chinese_chars=_A , strip_accents=_A , **_A , ) __snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( pre_tok_state.get('lowercase' , _A) != do_lower_case or pre_tok_state.get('strip_accents' , _A) != strip_accents ): __snake_case : Union[str, Any] = getattr(_A , pre_tok_state.pop('type')) __snake_case : Union[str, Any] = do_lower_case __snake_case : str = strip_accents __snake_case : Optional[int] = pre_tok_class(**_A) __snake_case : int = do_lower_case def __getstate__(self : Optional[Any]) -> Dict: __snake_case : Optional[int] = self.__dict__.copy() __snake_case : int = BertPreTokenizer() return state def __setstate__(self : Optional[Any] , _A : Optional[Any]) -> Dict: __snake_case : List[str] = d __snake_case : str = self.__dict__['_tokenizer'].get_vocab() __snake_case : int = PreTokenizer.custom(JiebaPreTokenizer(_A)) def _lowercase (self : int , _A : Tuple , _A : Any=None) -> str: __snake_case : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase (self : List[str] , _A : List[int] , _A : Optional[List[int]] = None) -> List[int]: __snake_case : Tuple = [self.sep_token_id] __snake_case : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _lowercase (self : List[Any] , _A : str , _A : Optional[str] = None) -> Tuple[str]: __snake_case : List[Any] = self._tokenizer.model.save(_A , name=_A) return tuple(_A) def _lowercase (self : int , _A : Optional[int] , _A : Tuple=None , _A : Tuple=None , _A : Dict=False , **_A : Optional[int] , ) -> Optional[Any]: __snake_case : Optional[Any] = BertPreTokenizer() return super().save_pretrained(_A , _A , _A , _A , **_A)
95
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case ( __snake_case, __snake_case, unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = StableDiffusionXLImgaImgPipeline SCREAMING_SNAKE_CASE_ : Dict = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width"""} SCREAMING_SNAKE_CASE_ : Optional[int] = PipelineTesterMixin.required_optional_params - {"""latents"""} SCREAMING_SNAKE_CASE_ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS SCREAMING_SNAKE_CASE_ : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS SCREAMING_SNAKE_CASE_ : Tuple = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase_ ( self : Union[str, Any])-> List[Any]: '''simple docstring''' torch.manual_seed(0) __lowerCAmelCase: Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , attention_head_dim=(2, 4) , use_linear_projection=UpperCamelCase__ , addition_embed_type="text_time" , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) __lowerCAmelCase: List[str] = EulerDiscreteScheduler( beta_start=0.00085 , beta_end=0.012 , steps_offset=1 , beta_schedule="scaled_linear" , timestep_spacing="leading" , ) torch.manual_seed(0) __lowerCAmelCase: Union[str, Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0) __lowerCAmelCase: Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="gelu" , projection_dim=3_2 , ) __lowerCAmelCase: Optional[Any] = CLIPTextModel(UpperCamelCase__) __lowerCAmelCase: Optional[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=UpperCamelCase__) __lowerCAmelCase: Tuple = CLIPTextModelWithProjection(UpperCamelCase__) __lowerCAmelCase: Tuple = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" , local_files_only=UpperCamelCase__) __lowerCAmelCase: List[str] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_encoder_2": text_encoder_a, "tokenizer_2": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def lowercase_ ( self : Union[str, Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any]=0)-> Any: '''simple docstring''' __lowerCAmelCase: Union[str, Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(UpperCamelCase__)).to(UpperCamelCase__) __lowerCAmelCase: List[str] = image / 2 + 0.5 if str(UpperCamelCase__).startswith("mps"): __lowerCAmelCase: Optional[Any] = torch.manual_seed(UpperCamelCase__) else: __lowerCAmelCase: Optional[int] = torch.Generator(device=UpperCamelCase__).manual_seed(UpperCamelCase__) __lowerCAmelCase: List[Any] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 5.0, "output_type": "numpy", "strength": 0.75, } return inputs def lowercase_ ( self : Dict)-> Union[str, Any]: '''simple docstring''' __lowerCAmelCase: Union[str, Any] = "cpu" # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase: int = self.get_dummy_components() __lowerCAmelCase: Optional[int] = StableDiffusionXLImgaImgPipeline(**UpperCamelCase__) __lowerCAmelCase: int = sd_pipe.to(UpperCamelCase__) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__) __lowerCAmelCase: str = self.get_dummy_inputs(UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = sd_pipe(**UpperCamelCase__).images __lowerCAmelCase: Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) __lowerCAmelCase: Any = np.array([0.4656, 0.4840, 0.4439, 0.6698, 0.5574, 0.4524, 0.5799, 0.5943, 0.5165]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def lowercase_ ( self : Tuple)-> List[str]: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3e-3) def lowercase_ ( self : List[Any])-> List[str]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3) def lowercase_ ( self : Optional[int])-> Any: '''simple docstring''' pass def lowercase_ ( self : Dict)-> Optional[int]: '''simple docstring''' __lowerCAmelCase: List[str] = self.get_dummy_components() __lowerCAmelCase: Optional[int] = StableDiffusionXLImgaImgPipeline(**UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = sd_pipe.to(UpperCamelCase__) __lowerCAmelCase: Tuple = sd_pipe.to(UpperCamelCase__) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__) # forward without prompt embeds __lowerCAmelCase: Tuple = self.get_dummy_inputs(UpperCamelCase__) __lowerCAmelCase: Any = 3 * ["this is a negative prompt"] __lowerCAmelCase: Tuple = negative_prompt __lowerCAmelCase: Any = 3 * [inputs["prompt"]] __lowerCAmelCase: Optional[Any] = sd_pipe(**UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = output.images[0, -3:, -3:, -1] # forward with prompt embeds __lowerCAmelCase: str = self.get_dummy_inputs(UpperCamelCase__) __lowerCAmelCase: Tuple = 3 * ["this is a negative prompt"] __lowerCAmelCase: str = 3 * [inputs.pop("prompt")] ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ): Optional[Any] = sd_pipe.encode_prompt(UpperCamelCase__ , negative_prompt=UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = sd_pipe( **UpperCamelCase__ , prompt_embeds=UpperCamelCase__ , negative_prompt_embeds=UpperCamelCase__ , pooled_prompt_embeds=UpperCamelCase__ , negative_pooled_prompt_embeds=UpperCamelCase__ , ) __lowerCAmelCase: Optional[int] = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten()).max() < 1e-4 @slow @require_torch_gpu class snake_case ( unittest.TestCase ): def lowercase_ ( self : Optional[int])-> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self : Optional[Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : List[str]="cpu" , UpperCamelCase__ : str=torch.floataa , UpperCamelCase__ : Any=0)-> Union[str, Any]: '''simple docstring''' __lowerCAmelCase: Dict = torch.Generator(device=UpperCamelCase__).manual_seed(UpperCamelCase__) __lowerCAmelCase: List[Any] = np.random.RandomState(UpperCamelCase__).standard_normal((1, 4, 6_4, 6_4)) __lowerCAmelCase: int = torch.from_numpy(UpperCamelCase__).to(device=UpperCamelCase__ , dtype=UpperCamelCase__) __lowerCAmelCase: List[str] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def lowercase_ ( self : str)-> List[str]: '''simple docstring''' __lowerCAmelCase: List[str] = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-base") pipe.to(UpperCamelCase__) pipe.set_progress_bar_config(disable=UpperCamelCase__) __lowerCAmelCase: Tuple = self.get_inputs(UpperCamelCase__) __lowerCAmelCase: str = pipe(**UpperCamelCase__).images __lowerCAmelCase: Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowerCAmelCase: Dict = np.array([0.49493, 0.47896, 0.40798, 0.54214, 0.53212, 0.48202, 0.47656, 0.46329, 0.48506]) assert np.abs(image_slice - expected_slice).max() < 7e-3
217
"""simple docstring""" def a__ ( __SCREAMING_SNAKE_CASE ) -> int: __lowerCAmelCase: Optional[Any] = 1 for i in range(1 , num + 1 ): fact *= i return fact def a__ ( __SCREAMING_SNAKE_CASE ) -> int: __lowerCAmelCase: List[str] = 0 while number > 0: __lowerCAmelCase: Any = number % 1_0 sum_of_digits += last_digit __lowerCAmelCase: List[Any] = number // 1_0 # Removing the last_digit from the given number return sum_of_digits def a__ ( __SCREAMING_SNAKE_CASE = 1_0_0 ) -> int: __lowerCAmelCase: Tuple = factorial(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Optional[int] = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
217
1
'''simple docstring''' from __future__ import annotations import os from collections.abc import Mapping __A : Union[str, Any] = tuple[int, int] class __UpperCamelCase : def __init__( self :Dict ,_UpperCamelCase :set[int] ,_UpperCamelCase :Mapping[EdgeT, int] ): snake_case_ : set[int] = vertices snake_case_ : dict[EdgeT, int] = { (min(_UpperCamelCase ), max(_UpperCamelCase )): weight for edge, weight in edges.items() } def a__ ( self :Union[str, Any] ,_UpperCamelCase :EdgeT ,_UpperCamelCase :int ): self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) snake_case_ : Optional[Any] = weight def a__ ( self :str ): snake_case_ : Graph = Graph({min(self.vertices )} ,{} ) snake_case_ : EdgeT snake_case_ : int snake_case_ : EdgeT snake_case_ : int while len(subgraph.vertices ) < len(self.vertices ): snake_case_ : Tuple = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: snake_case_ : Optional[Any] = edge snake_case_ : str = weight subgraph.add_edge(_UpperCamelCase ,_UpperCamelCase ) return subgraph def UpperCAmelCase ( lowerCamelCase_ :str = "p107_network.txt" ): '''simple docstring''' snake_case_ : str = os.path.abspath(os.path.dirname(lowerCamelCase_ ) ) snake_case_ : str = os.path.join(lowerCamelCase_ , lowerCamelCase_ ) snake_case_ : dict[EdgeT, int] = {} snake_case_ : list[str] snake_case_ : int snake_case_ : int with open(lowerCamelCase_ ) as f: snake_case_ : str = f.read().strip().split("""\n""" ) snake_case_ : Tuple = [line.split(""",""" ) for line in data] for edgea in range(1 , len(lowerCamelCase_ ) ): for edgea in range(lowerCamelCase_ ): if adjaceny_matrix[edgea][edgea] != "-": snake_case_ : str = int(adjaceny_matrix[edgea][edgea] ) snake_case_ : Graph = Graph(set(range(len(lowerCamelCase_ ) ) ) , lowerCamelCase_ ) snake_case_ : Graph = graph.prims_algorithm() snake_case_ : int = sum(graph.edges.values() ) snake_case_ : int = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F'{solution() = }')
8
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __UpperCamelCase ( unittest.TestCase ): def __init__( self :List[Any] ,_UpperCamelCase :List[str] ,_UpperCamelCase :Optional[Any]=7 ,_UpperCamelCase :Union[str, Any]=3 ,_UpperCamelCase :Any=1_8 ,_UpperCamelCase :Optional[Any]=3_0 ,_UpperCamelCase :List[str]=4_0_0 ,_UpperCamelCase :Optional[Any]=True ,_UpperCamelCase :Union[str, Any]=None ,_UpperCamelCase :List[Any]=True ,): snake_case_ : List[str] = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : str = batch_size snake_case_ : List[Any] = num_channels snake_case_ : Tuple = image_size snake_case_ : int = min_resolution snake_case_ : int = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[Any] = size snake_case_ : Any = apply_ocr def a__ ( self :Union[str, Any] ): return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class __UpperCamelCase ( lowercase__ , unittest.TestCase ): lowercase : Tuple = LayoutLMvaImageProcessor if is_pytesseract_available() else None def a__ ( self :List[Any] ): snake_case_ : Union[str, Any] = LayoutLMvaImageProcessingTester(self ) @property def a__ ( self :int ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self :Any ): snake_case_ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_UpperCamelCase ,"""do_resize""" ) ) self.assertTrue(hasattr(_UpperCamelCase ,"""size""" ) ) self.assertTrue(hasattr(_UpperCamelCase ,"""apply_ocr""" ) ) def a__ ( self :int ): snake_case_ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 1_8, """width""": 1_8} ) snake_case_ : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ) self.assertEqual(image_processor.size ,{"""height""": 4_2, """width""": 4_2} ) def a__ ( self :Optional[Any] ): pass def a__ ( self :Union[str, Any] ): # Initialize image_processing snake_case_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : List[str] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase ,Image.Image ) # Test not batched input snake_case_ : List[str] = image_processing(image_inputs[0] ,return_tensors="""pt""" ) self.assertEqual( encoding.pixel_values.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) self.assertIsInstance(encoding.words ,_UpperCamelCase ) self.assertIsInstance(encoding.boxes ,_UpperCamelCase ) # Test batched snake_case_ : List[Any] = image_processing(_UpperCamelCase ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def a__ ( self :Tuple ): # Initialize image_processing snake_case_ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Optional[Any] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCamelCase ,numpify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase ,np.ndarray ) # Test not batched input snake_case_ : Optional[int] = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched snake_case_ : Any = image_processing(_UpperCamelCase ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def a__ ( self :Optional[Any] ): # Initialize image_processing snake_case_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Optional[int] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=_UpperCamelCase ,torchify=_UpperCamelCase ) for image in image_inputs: self.assertIsInstance(_UpperCamelCase ,torch.Tensor ) # Test not batched input snake_case_ : Tuple = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched snake_case_ : Union[str, Any] = image_processing(_UpperCamelCase ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def a__ ( self :List[Any] ): # with apply_OCR = True snake_case_ : Any = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case_ : List[Any] = load_dataset("""hf-internal-testing/fixtures_docvqa""" ,split="""test""" ) snake_case_ : str = Image.open(ds[0]["""file"""] ).convert("""RGB""" ) snake_case_ : Dict = image_processing(_UpperCamelCase ,return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 2_2_4, 2_2_4) ) self.assertEqual(len(encoding.words ) ,len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case_ : Tuple = [["""11:14""", """to""", """11:39""", """a.m""", """11:39""", """to""", """11:44""", """a.m.""", """11:44""", """a.m.""", """to""", """12:25""", """p.m.""", """12:25""", """to""", """12:58""", """p.m.""", """12:58""", """to""", """4:00""", """p.m.""", """2:00""", """to""", """5:00""", """p.m.""", """Coffee""", """Break""", """Coffee""", """will""", """be""", """served""", """for""", """men""", """and""", """women""", """in""", """the""", """lobby""", """adjacent""", """to""", """exhibit""", """area.""", """Please""", """move""", """into""", """exhibit""", """area.""", """(Exhibits""", """Open)""", """TRRF""", """GENERAL""", """SESSION""", """(PART""", """|)""", """Presiding:""", """Lee""", """A.""", """Waller""", """TRRF""", """Vice""", """President""", """“Introductory""", """Remarks”""", """Lee""", """A.""", """Waller,""", """TRRF""", """Vice""", """Presi-""", """dent""", """Individual""", """Interviews""", """with""", """TRRF""", """Public""", """Board""", """Members""", """and""", """Sci-""", """entific""", """Advisory""", """Council""", """Mem-""", """bers""", """Conducted""", """by""", """TRRF""", """Treasurer""", """Philip""", """G.""", """Kuehn""", """to""", """get""", """answers""", """which""", """the""", """public""", """refrigerated""", """warehousing""", """industry""", """is""", """looking""", """for.""", """Plus""", """questions""", """from""", """the""", """floor.""", """Dr.""", """Emil""", """M.""", """Mrak,""", """University""", """of""", """Cal-""", """ifornia,""", """Chairman,""", """TRRF""", """Board;""", """Sam""", """R.""", """Cecil,""", """University""", """of""", """Georgia""", """College""", """of""", """Agriculture;""", """Dr.""", """Stanley""", """Charm,""", """Tufts""", """University""", """School""", """of""", """Medicine;""", """Dr.""", """Robert""", """H.""", """Cotton,""", """ITT""", """Continental""", """Baking""", """Company;""", """Dr.""", """Owen""", """Fennema,""", """University""", """of""", """Wis-""", """consin;""", """Dr.""", """Robert""", """E.""", """Hardenburg,""", """USDA.""", """Questions""", """and""", """Answers""", """Exhibits""", """Open""", """Capt.""", """Jack""", """Stoney""", """Room""", """TRRF""", """Scientific""", """Advisory""", """Council""", """Meeting""", """Ballroom""", """Foyer"""]] # noqa: E231 snake_case_ : Any = [[[1_4_1, 5_7, 2_1_4, 6_9], [2_2_8, 5_8, 2_5_2, 6_9], [1_4_1, 7_5, 2_1_6, 8_8], [2_3_0, 7_9, 2_8_0, 8_8], [1_4_2, 2_6_0, 2_1_8, 2_7_3], [2_3_0, 2_6_1, 2_5_5, 2_7_3], [1_4_3, 2_7_9, 2_1_8, 2_9_0], [2_3_1, 2_8_2, 2_9_0, 2_9_1], [1_4_3, 3_4_2, 2_1_8, 3_5_4], [2_3_1, 3_4_5, 2_8_9, 3_5_5], [2_0_2, 3_6_2, 2_2_7, 3_7_3], [1_4_3, 3_7_9, 2_2_0, 3_9_2], [2_3_1, 3_8_2, 2_9_1, 3_9_4], [1_4_4, 7_1_4, 2_2_0, 7_2_6], [2_3_1, 7_1_5, 2_5_6, 7_2_6], [1_4_4, 7_3_2, 2_2_0, 7_4_5], [2_3_2, 7_3_6, 2_9_1, 7_4_7], [1_4_4, 7_6_9, 2_1_8, 7_8_2], [2_3_1, 7_7_0, 2_5_6, 7_8_2], [1_4_1, 7_8_8, 2_0_2, 8_0_1], [2_1_5, 7_9_1, 2_7_4, 8_0_4], [1_4_3, 8_2_6, 2_0_4, 8_3_8], [2_1_5, 8_2_6, 2_4_0, 8_3_8], [1_4_2, 8_4_4, 2_0_2, 8_5_7], [2_1_5, 8_4_7, 2_7_4, 8_5_9], [3_3_4, 5_7, 4_2_7, 6_9], [4_4_0, 5_7, 5_2_2, 6_9], [3_6_9, 7_5, 4_6_1, 8_8], [4_6_9, 7_5, 5_1_6, 8_8], [5_2_8, 7_6, 5_6_2, 8_8], [5_7_0, 7_6, 6_6_7, 8_8], [6_7_5, 7_5, 7_1_1, 8_7], [7_2_1, 7_9, 7_7_8, 8_8], [7_8_9, 7_5, 8_4_0, 8_8], [3_6_9, 9_7, 4_7_0, 1_0_7], [4_8_4, 9_4, 5_0_7, 1_0_6], [5_1_8, 9_4, 5_6_2, 1_0_7], [5_7_6, 9_4, 6_5_5, 1_1_0], [6_6_8, 9_4, 7_9_2, 1_0_9], [8_0_4, 9_5, 8_2_9, 1_0_7], [3_6_9, 1_1_3, 4_6_5, 1_2_5], [4_7_7, 1_1_6, 5_4_7, 1_2_5], [5_6_2, 1_1_3, 6_5_8, 1_2_5], [6_7_1, 1_1_6, 7_4_8, 1_2_5], [7_6_1, 1_1_3, 8_1_1, 1_2_5], [3_6_9, 1_3_1, 4_6_5, 1_4_3], [4_7_7, 1_3_3, 5_4_8, 1_4_3], [5_6_3, 1_3_0, 6_9_8, 1_4_5], [7_1_0, 1_3_0, 8_0_2, 1_4_6], [3_3_6, 1_7_1, 4_1_2, 1_8_3], [4_2_3, 1_7_1, 5_7_2, 1_8_3], [5_8_2, 1_7_0, 7_1_6, 1_8_4], [7_2_8, 1_7_1, 8_1_7, 1_8_7], [8_2_9, 1_7_1, 8_4_4, 1_8_6], [3_3_8, 1_9_7, 4_8_2, 2_1_2], [5_0_7, 1_9_6, 5_5_7, 2_0_9], [5_6_9, 1_9_6, 5_9_5, 2_0_8], [6_1_0, 1_9_6, 7_0_2, 2_0_9], [5_0_5, 2_1_4, 5_8_3, 2_2_6], [5_9_5, 2_1_4, 6_5_6, 2_2_7], [6_7_0, 2_1_5, 8_0_7, 2_2_7], [3_3_5, 2_5_9, 5_4_3, 2_7_4], [5_5_6, 2_5_9, 7_0_8, 2_7_2], [3_7_2, 2_7_9, 4_2_2, 2_9_1], [4_3_5, 2_7_9, 4_6_0, 2_9_1], [4_7_4, 2_7_9, 5_7_4, 2_9_2], [5_8_7, 2_7_8, 6_6_4, 2_9_1], [6_7_6, 2_7_8, 7_3_8, 2_9_1], [7_5_1, 2_7_9, 8_3_4, 2_9_1], [3_7_2, 2_9_8, 4_3_4, 3_1_0], [3_3_5, 3_4_1, 4_8_3, 3_5_4], [4_9_7, 3_4_1, 6_5_5, 3_5_4], [6_6_7, 3_4_1, 7_2_8, 3_5_4], [7_4_0, 3_4_1, 8_2_5, 3_5_4], [3_3_5, 3_6_0, 4_3_0, 3_7_2], [4_4_2, 3_6_0, 5_3_4, 3_7_2], [5_4_5, 3_5_9, 6_8_7, 3_7_2], [6_9_7, 3_6_0, 7_5_4, 3_7_2], [7_6_5, 3_6_0, 8_2_3, 3_7_3], [3_3_4, 3_7_8, 4_2_8, 3_9_1], [4_4_0, 3_7_8, 5_7_7, 3_9_4], [5_9_0, 3_7_8, 7_0_5, 3_9_1], [7_2_0, 3_7_8, 8_0_1, 3_9_1], [3_3_4, 3_9_7, 4_0_0, 4_0_9], [3_7_0, 4_1_6, 5_2_9, 4_2_9], [5_4_4, 4_1_6, 5_7_6, 4_3_2], [5_8_7, 4_1_6, 6_6_5, 4_2_8], [6_7_7, 4_1_6, 8_1_4, 4_2_9], [3_7_2, 4_3_5, 4_5_2, 4_5_0], [4_6_5, 4_3_4, 4_9_5, 4_4_7], [5_1_1, 4_3_4, 6_0_0, 4_4_7], [6_1_1, 4_3_6, 6_3_7, 4_4_7], [6_4_9, 4_3_6, 6_9_4, 4_5_1], [7_0_5, 4_3_8, 8_2_4, 4_4_7], [3_6_9, 4_5_3, 4_5_2, 4_6_6], [4_6_4, 4_5_4, 5_0_9, 4_6_6], [5_2_2, 4_5_3, 6_1_1, 4_6_9], [6_2_5, 4_5_3, 7_9_2, 4_6_9], [3_7_0, 4_7_2, 5_5_6, 4_8_8], [5_7_0, 4_7_2, 6_8_4, 4_8_7], [6_9_7, 4_7_2, 7_1_8, 4_8_5], [7_3_2, 4_7_2, 8_3_5, 4_8_8], [3_6_9, 4_9_0, 4_1_1, 5_0_3], [4_2_5, 4_9_0, 4_8_4, 5_0_3], [4_9_6, 4_9_0, 6_3_5, 5_0_6], [6_4_5, 4_9_0, 7_0_7, 5_0_3], [7_1_8, 4_9_1, 7_6_1, 5_0_3], [7_7_1, 4_9_0, 8_4_0, 5_0_3], [3_3_6, 5_1_0, 3_7_4, 5_2_1], [3_8_8, 5_1_0, 4_4_7, 5_2_2], [4_6_0, 5_1_0, 4_8_9, 5_2_1], [5_0_3, 5_1_0, 5_8_0, 5_2_2], [5_9_2, 5_0_9, 7_3_6, 5_2_5], [7_4_5, 5_0_9, 7_7_0, 5_2_2], [7_8_1, 5_0_9, 8_4_0, 5_2_2], [3_3_8, 5_2_8, 4_3_4, 5_4_1], [4_4_8, 5_2_8, 5_9_6, 5_4_1], [6_0_9, 5_2_7, 6_8_7, 5_4_0], [7_0_0, 5_2_8, 7_9_2, 5_4_1], [3_3_6, 5_4_6, 3_9_7, 5_5_9], [4_0_7, 5_4_6, 4_3_1, 5_5_9], [4_4_3, 5_4_6, 5_2_5, 5_6_0], [5_3_7, 5_4_6, 6_8_0, 5_6_2], [6_8_8, 5_4_6, 7_1_4, 5_5_9], [7_2_2, 5_4_6, 8_3_7, 5_6_2], [3_3_6, 5_6_5, 4_4_9, 5_8_1], [4_6_1, 5_6_5, 4_8_5, 5_7_7], [4_9_7, 5_6_5, 6_6_5, 5_8_1], [6_8_1, 5_6_5, 7_1_8, 5_7_7], [7_3_2, 5_6_5, 8_3_7, 5_8_0], [3_3_7, 5_8_4, 4_3_8, 5_9_7], [4_5_2, 5_8_3, 5_2_1, 5_9_6], [5_3_5, 5_8_4, 6_7_7, 5_9_9], [6_9_0, 5_8_3, 7_8_7, 5_9_6], [8_0_1, 5_8_3, 8_2_5, 5_9_6], [3_3_8, 6_0_2, 4_7_8, 6_1_5], [4_9_2, 6_0_2, 5_3_0, 6_1_4], [5_4_3, 6_0_2, 6_3_8, 6_1_5], [6_5_0, 6_0_2, 6_7_6, 6_1_4], [6_8_8, 6_0_2, 7_8_8, 6_1_5], [8_0_2, 6_0_2, 8_4_3, 6_1_4], [3_3_7, 6_2_1, 5_0_2, 6_3_3], [5_1_6, 6_2_1, 6_1_5, 6_3_7], [6_2_9, 6_2_1, 7_7_4, 6_3_6], [7_8_9, 6_2_1, 8_2_7, 6_3_3], [3_3_7, 6_3_9, 4_1_8, 6_5_2], [4_3_2, 6_4_0, 5_7_1, 6_5_3], [5_8_7, 6_3_9, 7_3_1, 6_5_5], [7_4_3, 6_3_9, 7_6_9, 6_5_2], [7_8_0, 6_3_9, 8_4_1, 6_5_2], [3_3_8, 6_5_8, 4_4_0, 6_7_3], [4_5_5, 6_5_8, 4_9_1, 6_7_0], [5_0_8, 6_5_8, 6_0_2, 6_7_1], [6_1_6, 6_5_8, 6_3_8, 6_7_0], [6_5_4, 6_5_8, 8_3_5, 6_7_4], [3_3_7, 6_7_7, 4_2_9, 6_8_9], [3_3_7, 7_1_4, 4_8_2, 7_2_6], [4_9_5, 7_1_4, 5_4_8, 7_2_6], [5_6_1, 7_1_4, 6_8_3, 7_2_6], [3_3_8, 7_7_0, 4_6_1, 7_8_2], [4_7_4, 7_6_9, 5_5_4, 7_8_5], [4_8_9, 7_8_8, 5_6_2, 8_0_3], [5_7_6, 7_8_8, 6_4_3, 8_0_1], [6_5_6, 7_8_7, 7_5_1, 8_0_4], [7_6_4, 7_8_8, 8_4_4, 8_0_1], [3_3_4, 8_2_5, 4_2_1, 8_3_8], [4_3_0, 8_2_4, 5_7_4, 8_3_8], [5_8_4, 8_2_4, 7_2_3, 8_4_1], [3_3_5, 8_4_4, 4_5_0, 8_5_7], [4_6_4, 8_4_3, 5_8_3, 8_6_0], [6_2_8, 8_6_2, 7_5_5, 8_7_5], [7_6_9, 8_6_1, 8_4_8, 8_7_8]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words ,_UpperCamelCase ) self.assertListEqual(encoding.boxes ,_UpperCamelCase ) # with apply_OCR = False snake_case_ : Dict = LayoutLMvaImageProcessor(apply_ocr=_UpperCamelCase ) snake_case_ : Optional[int] = image_processing(_UpperCamelCase ,return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 2_2_4, 2_2_4) )
8
1
'''simple docstring''' import argparse import datetime def __lowerCamelCase ( A__ ) -> str: """simple docstring""" UpperCamelCase = { '0': 'Sunday', '1': 'Monday', '2': 'Tuesday', '3': 'Wednesday', '4': 'Thursday', '5': 'Friday', '6': 'Saturday', } UpperCamelCase = {0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(A__ ) < 11: raise ValueError('Must be 10 characters long' ) # Get month UpperCamelCase = int(date_input[0] + date_input[1] ) # Validate if not 0 < m < 13: raise ValueError('Month must be between 1 - 12' ) UpperCamelCase = date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError('Date separator must be \'-\' or \'/\'' ) # Get day UpperCamelCase = int(date_input[3] + date_input[4] ) # Validate if not 0 < d < 32: raise ValueError('Date must be between 1 - 31' ) # Get second separator UpperCamelCase = date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError('Date separator must be \'-\' or \'/\'' ) # Get year UpperCamelCase = int(date_input[6] + date_input[7] + date_input[8] + date_input[9] ) # Arbitrary year range if not 45 < y < 8_500: raise ValueError( 'Year out of range. There has to be some sort of limit...right?' ) # Get datetime obj for validation UpperCamelCase = datetime.date(int(A__ ) , int(A__ ) , int(A__ ) ) # Start math if m <= 2: UpperCamelCase = y - 1 UpperCamelCase = m + 12 # maths var UpperCamelCase = int(str(A__ )[:2] ) UpperCamelCase = int(str(A__ )[2:] ) UpperCamelCase = int(2.6 * m - 5.39 ) UpperCamelCase = int(c / 4 ) UpperCamelCase = int(k / 4 ) UpperCamelCase = int(d + k ) UpperCamelCase = int(t + u + v + x ) UpperCamelCase = int(z - (2 * c) ) UpperCamelCase = round(w % 7 ) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError('The date was evaluated incorrectly. Contact developer.' ) # Response UpperCamelCase = F"""Your date {date_input}, is a {days[str(A__ )]}!""" return response if __name__ == "__main__": import doctest doctest.testmod() _lowerCamelCase : Optional[int] = argparse.ArgumentParser( description=( "Find out what day of the week nearly any date is or was. Enter " "date as a string in the mm-dd-yyyy or mm/dd/yyyy format" ) ) parser.add_argument( "date_input", type=str, help="Date as a string (mm-dd-yyyy or mm/dd/yyyy)" ) _lowerCamelCase : Optional[Any] = parser.parse_args() zeller(args.date_input)
28
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ : Dict = logging.get_logger(__name__) UpperCAmelCase__ : str = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class a__ ( UpperCAmelCase ): """simple docstring""" UpperCAmelCase__ : List[Any] ="""nllb-moe""" UpperCAmelCase__ : Any =["""past_key_values"""] UpperCAmelCase__ : Dict ={"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[Any]=1_2_8_1_1_2 , UpperCAmelCase__ : Tuple=1_0_2_4 , UpperCAmelCase__ : str=1_2 , UpperCAmelCase__ : int=4_0_9_6 , UpperCAmelCase__ : Dict=1_6 , UpperCAmelCase__ : Union[str, Any]=1_2 , UpperCAmelCase__ : int=4_0_9_6 , UpperCAmelCase__ : Optional[Any]=1_6 , UpperCAmelCase__ : Union[str, Any]=0.05 , UpperCAmelCase__ : Any=0.05 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Union[str, Any]="relu" , UpperCAmelCase__ : Dict=1_0_2_4 , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : str=0.1 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : Tuple=0.02 , UpperCAmelCase__ : Optional[int]=2 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Dict="float32" , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Union[str, Any]=1_2_8 , UpperCAmelCase__ : Any=6_4 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : List[Any]=4 , UpperCAmelCase__ : Optional[Any]=0.0_01 , UpperCAmelCase__ : Optional[Any]=0.0_01 , UpperCAmelCase__ : Dict="all" , UpperCAmelCase__ : Optional[int]=False , UpperCAmelCase__ : List[Any]=False , UpperCAmelCase__ : List[str]=1.0 , UpperCAmelCase__ : Optional[int]=0.2 , UpperCAmelCase__ : Dict=1 , UpperCAmelCase__ : List[Any]=0 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Tuple=False , **UpperCAmelCase__ : Union[str, Any] , ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = vocab_size SCREAMING_SNAKE_CASE : Optional[int] = max_position_embeddings SCREAMING_SNAKE_CASE : str = d_model SCREAMING_SNAKE_CASE : Dict = encoder_ffn_dim SCREAMING_SNAKE_CASE : Union[str, Any] = encoder_layers SCREAMING_SNAKE_CASE : str = encoder_attention_heads SCREAMING_SNAKE_CASE : List[str] = decoder_ffn_dim SCREAMING_SNAKE_CASE : str = decoder_layers SCREAMING_SNAKE_CASE : Optional[Any] = decoder_attention_heads SCREAMING_SNAKE_CASE : Optional[Any] = dropout SCREAMING_SNAKE_CASE : Optional[int] = attention_dropout SCREAMING_SNAKE_CASE : str = activation_dropout SCREAMING_SNAKE_CASE : Dict = activation_function SCREAMING_SNAKE_CASE : str = init_std SCREAMING_SNAKE_CASE : Tuple = encoder_layerdrop SCREAMING_SNAKE_CASE : Tuple = decoder_layerdrop SCREAMING_SNAKE_CASE : List[str] = use_cache SCREAMING_SNAKE_CASE : Optional[Any] = encoder_layers SCREAMING_SNAKE_CASE : List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True SCREAMING_SNAKE_CASE : Tuple = router_z_loss_coef SCREAMING_SNAKE_CASE : Tuple = router_aux_loss_coef SCREAMING_SNAKE_CASE : List[Any] = decoder_sparse_step SCREAMING_SNAKE_CASE : Any = encoder_sparse_step SCREAMING_SNAKE_CASE : Tuple = num_experts SCREAMING_SNAKE_CASE : Optional[int] = expert_capacity SCREAMING_SNAKE_CASE : int = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) SCREAMING_SNAKE_CASE : Optional[int] = router_dtype SCREAMING_SNAKE_CASE : Any = router_ignore_padding_tokens SCREAMING_SNAKE_CASE : Any = batch_prioritized_routing SCREAMING_SNAKE_CASE : Optional[Any] = second_expert_policy SCREAMING_SNAKE_CASE : Any = normalize_router_prob_before_dropping SCREAMING_SNAKE_CASE : Tuple = moe_eval_capacity_token_fraction SCREAMING_SNAKE_CASE : int = moe_token_dropout SCREAMING_SNAKE_CASE : Optional[int] = output_router_logits super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , )
245
0
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class A__ : '''simple docstring''' SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 SCREAMING_SNAKE_CASE = 42 def _SCREAMING_SNAKE_CASE ( self: Any) -> List[str]: """simple docstring""" assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape) == len(self.y.shape) == len(self.z.shape) == len(self.origin.shape) == 2 def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Tuple: """simple docstring""" return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa)) def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> List[str]: """simple docstring""" return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa)) def _SCREAMING_SNAKE_CASE ( self: int) -> torch.Tensor: """simple docstring""" __lowerCAmelCase : List[Any] = torch.arange(self.height * self.width) __lowerCAmelCase : Union[str, Any] = torch.stack( [ pixel_indices % self.width, torch.div(_SCREAMING_SNAKE_CASE , self.width , rounding_mode="trunc"), ] , axis=1 , ) return coords @property def _SCREAMING_SNAKE_CASE ( self: List[str]) -> Optional[Any]: """simple docstring""" __lowerCAmelCase , *__lowerCAmelCase : Dict = self.shape __lowerCAmelCase : Any = int(np.prod(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : int = self.get_image_coords() __lowerCAmelCase : List[Any] = torch.broadcast_to(coords.unsqueeze(0) , [batch_size * inner_batch_size, *coords.shape]) __lowerCAmelCase : Optional[int] = self.get_camera_rays(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Tuple = rays.view(_SCREAMING_SNAKE_CASE , inner_batch_size * self.height * self.width , 2 , 3) return rays def _SCREAMING_SNAKE_CASE ( self: str , _SCREAMING_SNAKE_CASE: torch.Tensor) -> torch.Tensor: """simple docstring""" __lowerCAmelCase , *__lowerCAmelCase , __lowerCAmelCase : Optional[Any] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __lowerCAmelCase : str = coords.view(_SCREAMING_SNAKE_CASE , -1 , 2) __lowerCAmelCase : Union[str, Any] = self.resolution() __lowerCAmelCase : Tuple = self.fov() __lowerCAmelCase : List[str] = (flat.float() / (res - 1)) * 2 - 1 __lowerCAmelCase : Any = fracs * torch.tan(fov / 2) __lowerCAmelCase : Dict = fracs.view(_SCREAMING_SNAKE_CASE , -1 , 2) __lowerCAmelCase : Dict = ( self.z.view(_SCREAMING_SNAKE_CASE , 1 , 3) + self.x.view(_SCREAMING_SNAKE_CASE , 1 , 3) * fracs[:, :, :1] + self.y.view(_SCREAMING_SNAKE_CASE , 1 , 3) * fracs[:, :, 1:] ) __lowerCAmelCase : int = directions / directions.norm(dim=-1 , keepdim=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : int = torch.stack( [ torch.broadcast_to(self.origin.view(_SCREAMING_SNAKE_CASE , 1 , 3) , [batch_size, directions.shape[1], 3]), directions, ] , dim=2 , ) return rays.view(_SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , 2 , 3) def _SCREAMING_SNAKE_CASE ( self: Any , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: int) -> "DifferentiableProjectiveCamera": """simple docstring""" assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=_SCREAMING_SNAKE_CASE , height=_SCREAMING_SNAKE_CASE , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowercase ( __snake_case ) -> DifferentiableProjectiveCamera: __lowerCAmelCase : Any = [] __lowerCAmelCase : Optional[Any] = [] __lowerCAmelCase : List[str] = [] __lowerCAmelCase : Optional[int] = [] for theta in np.linspace(0 ,2 * np.pi ,num=20 ): __lowerCAmelCase : List[str] = np.array([np.sin(__snake_case ), np.cos(__snake_case ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __lowerCAmelCase : int = -z * 4 __lowerCAmelCase : int = np.array([np.cos(__snake_case ), -np.sin(__snake_case ), 0.0] ) __lowerCAmelCase : Any = np.cross(__snake_case ,__snake_case ) origins.append(__snake_case ) xs.append(__snake_case ) ys.append(__snake_case ) zs.append(__snake_case ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(__snake_case ,axis=0 ) ).float() ,x=torch.from_numpy(np.stack(__snake_case ,axis=0 ) ).float() ,y=torch.from_numpy(np.stack(__snake_case ,axis=0 ) ).float() ,z=torch.from_numpy(np.stack(__snake_case ,axis=0 ) ).float() ,width=__snake_case ,height=__snake_case ,x_fov=0.7 ,y_fov=0.7 ,shape=(1, len(__snake_case )) ,)
58
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
58
1
'''simple docstring''' _lowerCAmelCase = frozenset( [ '''prompt''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', '''cross_attention_kwargs''', ] ) _lowerCAmelCase = frozenset(['''prompt''', '''negative_prompt''']) _lowerCAmelCase = frozenset([]) _lowerCAmelCase = frozenset(['''image''']) _lowerCAmelCase = frozenset( [ '''image''', '''height''', '''width''', '''guidance_scale''', ] ) _lowerCAmelCase = frozenset(['''image''']) _lowerCAmelCase = frozenset( [ '''prompt''', '''image''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', ] ) _lowerCAmelCase = frozenset(['''prompt''', '''image''', '''negative_prompt''']) _lowerCAmelCase = frozenset( [ # Text guided image variation with an image mask '''prompt''', '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', ] ) _lowerCAmelCase = frozenset(['''prompt''', '''image''', '''mask_image''', '''negative_prompt''']) _lowerCAmelCase = frozenset( [ # image variation with an image mask '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', ] ) _lowerCAmelCase = frozenset(['''image''', '''mask_image''']) _lowerCAmelCase = frozenset( [ '''example_image''', '''image''', '''mask_image''', '''height''', '''width''', '''guidance_scale''', ] ) _lowerCAmelCase = frozenset(['''example_image''', '''image''', '''mask_image''']) _lowerCAmelCase = frozenset(['''class_labels''']) _lowerCAmelCase = frozenset(['''class_labels''']) _lowerCAmelCase = frozenset(['''batch_size''']) _lowerCAmelCase = frozenset([]) _lowerCAmelCase = frozenset(['''batch_size''']) _lowerCAmelCase = frozenset([]) _lowerCAmelCase = frozenset( [ '''prompt''', '''audio_length_in_s''', '''guidance_scale''', '''negative_prompt''', '''prompt_embeds''', '''negative_prompt_embeds''', '''cross_attention_kwargs''', ] ) _lowerCAmelCase = frozenset(['''prompt''', '''negative_prompt''']) _lowerCAmelCase = frozenset(['''input_tokens''']) _lowerCAmelCase = frozenset(['''input_tokens'''])
37
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline __snake_case = argparse.ArgumentParser('''Stable Diffusion script with intel optimization''', add_help=False) parser.add_argument('''--dpm''', action='''store_true''', help='''Enable DPMSolver or not''') parser.add_argument('''--steps''', default=None, type=int, help='''Num inference steps''') __snake_case = parser.parse_args() __snake_case = '''cpu''' __snake_case = '''a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings''' __snake_case = '''path-to-your-trained-model''' __snake_case = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: __snake_case = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __snake_case = pipe.to(device) # to channels last __snake_case = pipe.unet.to(memory_format=torch.channels_last) __snake_case = pipe.vae.to(memory_format=torch.channels_last) __snake_case = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: __snake_case = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex __snake_case = torch.randn(2, 4, 64, 64) __snake_case = torch.rand(1) * 9_99 __snake_case = torch.randn(2, 77, 7_68) __snake_case = (sample, timestep, encoder_hidden_status) try: __snake_case = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: __snake_case = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) __snake_case = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) __snake_case = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: __snake_case = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute __snake_case = 6_66 __snake_case = torch.Generator(device).manual_seed(seed) __snake_case = {'''generator''': generator} if args.steps is not None: __snake_case = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): __snake_case = pipe(prompt, **generate_kwargs).images[0] # save image image.save('''generated.png''')
348
0
'''simple docstring''' import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __a: Any = logging.get_logger(__name__) __a: Tuple = {"""vocab_file""": """vocab.txt""", """emoji_file""": """emoji.json"""} __a: Optional[int] = { """vocab_file""": { """abeja/gpt-neox-japanese-2.7b""": """https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt""", }, """emoji_file""": { """abeja/gpt-neox-japanese-2.7b""": """https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json""", }, } __a: Union[str, Any] = { """abeja/gpt-neox-japanese-2.7b""": 20_48, } def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase ): with open(UpperCAmelCase , '''r''' , encoding='''utf-8''' ) as f: lowercase__ : str = json.loads(f.read() ) lowercase__ : Any = collections.OrderedDict() lowercase__ : Optional[Any] = collections.OrderedDict() lowercase__ : Optional[int] = collections.OrderedDict() with open(UpperCAmelCase , '''r''' , encoding='''utf-8''' ) as f: lowercase__ : Optional[int] = f.readlines() lowercase__ : Union[str, Any] = [[t.rstrip('''\n''' )] if (t == ''',''' or ''',''' not in t) else t.rstrip('''\n''' ).split(''',''' ) for t in token] for idx, b in enumerate(UpperCAmelCase ): lowercase__ : Optional[Any] = b lowercase__ : Union[str, Any] = idx for wd in b: lowercase__ : List[Any] = idx return vocab, raw_vocab, ids_to_tokens, emoji class UpperCAmelCase ( a__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = ["input_ids", "attention_mask"] def __init__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase="<|endoftext|>" , __lowerCAmelCase="<|endoftext|>" , __lowerCAmelCase="<|startoftext|>" , __lowerCAmelCase="<|endoftext|>" , __lowerCAmelCase=False , **__lowerCAmelCase , ) -> List[str]: super().__init__( unk_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , do_clean_text=__lowerCAmelCase , **__lowerCAmelCase , ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError( F"""Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained""" ''' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`''' ) if not os.path.isfile(__lowerCAmelCase ): raise ValueError( F"""Can't find a emoji file at path '{emoji_file}'. To load the emoji information from a Google""" ''' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`''' ) lowercase__ : Optional[Any] = do_clean_text lowercase__ : Optional[Any] = load_vocab_and_emoji(__lowerCAmelCase , __lowerCAmelCase ) lowercase__ : Tuple = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def _lowerCAmelCase( self ) -> Union[str, Any]: # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab ) def _lowerCAmelCase( self ) -> List[str]: return dict(self.raw_vocab , **self.added_tokens_encoder ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> Optional[Any]: return self.subword_tokenizer.tokenize(__lowerCAmelCase , clean=self.do_clean_text ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> List[Any]: return self.vocab.get(__lowerCAmelCase , self.vocab.get(self.unk_token ) ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> str: return self.subword_tokenizer.convert_id_to_token(__lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> Dict: lowercase__ : Optional[Any] = ''''''.join(__lowerCAmelCase ).strip() return out_string def _lowerCAmelCase( self , __lowerCAmelCase ) -> List[int]: lowercase__ : Union[str, Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase ) + [self.eos_token_id] ) if len(__lowerCAmelCase ) > self.model_max_length: lowercase__ : Optional[int] = input_ids[-self.model_max_length :] return input_ids def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase = None ) -> Tuple[str]: lowercase__ : Optional[Any] = 0 if os.path.isdir(__lowerCAmelCase ): lowercase__ : List[str] = os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase__ : Optional[Any] = os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''emoji_file'''] ) else: lowercase__ : List[str] = ( (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase__ : int = ( (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory + VOCAB_FILES_NAMES['''emoji_file'''] ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ''' Please check that the vocabulary is not corrupted!''' ) lowercase__ : List[Any] = token_index writer.write(''','''.join(__lowerCAmelCase ) + '''\n''' ) index += 1 with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as writer: json.dump(self.emoji , __lowerCAmelCase ) return vocab_file, emoji_file class UpperCAmelCase ( a__ ): '''simple docstring''' def __init__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: lowercase__ : Union[str, Any] = vocab # same as swe lowercase__ : List[str] = ids_to_tokens # same as bpe lowercase__ : str = emoji lowercase__ : str = np.max([len(__lowerCAmelCase ) for w in self.vocab.keys()] ) lowercase__ : Dict = re.compile(r'''(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)''' ) lowercase__ : Dict = re.compile(r'''[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*''' ) lowercase__ : Union[str, Any] = re.compile(r'''[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}''' ) lowercase__ : List[str] = re.compile( r'''([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*''' ) lowercase__ : Optional[int] = re.compile( r'''(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*''' ) lowercase__ : List[Any] = re.compile( r'''((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*''' ) lowercase__ : Any = '''─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿''' lowercase__ : Optional[int] = '''▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟''' lowercase__ : Union[str, Any] = str.maketrans({k: '''<BLOCK>''' for k in keisen + blocks} ) def __len__( self ) -> Optional[int]: return len(self.ids_to_tokens ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> Tuple: lowercase__ : Optional[int] = self.content_repattera.sub('''<URL>''' , __lowerCAmelCase ) lowercase__ : List[str] = self.content_repattera.sub('''<EMAIL>''' , __lowerCAmelCase ) lowercase__ : int = self.content_repattera.sub('''<TEL>''' , __lowerCAmelCase ) lowercase__ : Dict = self.content_repattera.sub('''<DATE>''' , __lowerCAmelCase ) lowercase__ : List[Any] = self.content_repattera.sub('''<DATE>''' , __lowerCAmelCase ) lowercase__ : Any = self.content_repattera.sub('''<PRICE>''' , __lowerCAmelCase ) lowercase__ : List[str] = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: lowercase__ : List[Any] = content.replace('''<BLOCK><BLOCK>''' , '''<BLOCK>''' ) return content def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase=False ) -> Union[str, Any]: lowercase__ : Optional[int] = text.replace(''' ''' , '''<SP>''' ) lowercase__ : Optional[int] = text.replace(''' ''' , '''<SP>''' ) lowercase__ : str = text.replace('''\r\n''' , '''<BR>''' ) lowercase__ : Dict = text.replace('''\n''' , '''<BR>''' ) lowercase__ : int = text.replace('''\r''' , '''<BR>''' ) lowercase__ : Any = text.replace('''\t''' , '''<TAB>''' ) lowercase__ : List[str] = text.replace('''—''' , '''ー''' ) lowercase__ : Tuple = text.replace('''−''' , '''ー''' ) for k, v in self.emoji["emoji"].items(): if k in text: lowercase__ : Any = text.replace(__lowerCAmelCase , __lowerCAmelCase ) if clean: lowercase__ : List[str] = self.clean_text(__lowerCAmelCase ) def check_simbol(__lowerCAmelCase ): lowercase__ : Tuple = x.encode() if len(__lowerCAmelCase ) == 1 and len(__lowerCAmelCase ) == 2: lowercase__ : Tuple = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0XC2A1 and c <= 0XC2BF) or (c >= 0XC780 and c <= 0XC783) or (c >= 0XCAB9 and c <= 0XCBBF) or (c >= 0XCC80 and c <= 0XCDA2) ): return True return False def checkuae(__lowerCAmelCase ): lowercase__ : Optional[int] = x.encode() if len(__lowerCAmelCase ) == 1 and len(__lowerCAmelCase ) == 3: lowercase__ : Any = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0XE28080 and c <= 0XE2B07F: return True return False lowercase__ : str = 0 lowercase__ : List[Any] = [] while pos < len(__lowerCAmelCase ): lowercase__ : Optional[int] = min(len(__lowerCAmelCase ) , pos + self.maxlen + 1 ) if text[pos] == '''<''' else pos + 3 lowercase__ : Optional[int] = [] # (token_id, token, pos) for e in range(__lowerCAmelCase , __lowerCAmelCase , -1 ): lowercase__ : Union[str, Any] = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(__lowerCAmelCase ) > 2: lowercase__ : Optional[Any] = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(__lowerCAmelCase ) > 0: # the smallest token_id is adopted lowercase__ : Optional[int] = sorted(__lowerCAmelCase , key=lambda __lowerCAmelCase : x[0] )[0] result.append(__lowerCAmelCase ) lowercase__ : int = e else: lowercase__ : Optional[Any] = pos + 1 lowercase__ : List[str] = text[pos:end] if check_simbol(__lowerCAmelCase ): result.append('''<KIGOU>''' ) elif checkuae(__lowerCAmelCase ): result.append('''<U2000U2BFF>''' ) else: for i in wd.encode('''utf-8''' ): result.append('''<|byte%d|>''' % i ) lowercase__ : int = end return result def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase="\n" ) -> str: lowercase__ : Union[str, Any] = [] lowercase__ : int = [] lowercase__ : str = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(__lowerCAmelCase ) > 0: words.append(bytearray(__lowerCAmelCase ).decode('''utf-8''' , errors='''replace''' ) ) lowercase__ : List[Any] = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['''emoji_inv'''][word] ) elif word == "<SP>": words.append(''' ''' ) elif word == "<BR>": words.append(__lowerCAmelCase ) elif word == "<TAB>": words.append('''\t''' ) elif word == "<BLOCK>": words.append('''▀''' ) elif word == "<KIGOU>": words.append('''ǀ''' ) elif word == "<U2000U2BFF>": words.append('''‖''' ) else: words.append(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: words.append(bytearray(__lowerCAmelCase ).decode('''utf-8''' , errors='''replace''' ) ) lowercase__ : Dict = ''''''.join(__lowerCAmelCase ) return text
369
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase ): if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) lowercase__ : Tuple = sorted(string.lower() ) return len(UpperCAmelCase ) == len(set(UpperCAmelCase ) ) if __name__ == "__main__": __a: Union[str, Any] = input("""Enter a string """).strip() __a: Tuple = is_isogram(input_str) print(F'{input_str} is {"an" if isogram else "not an"} isogram.')
214
0
from typing import Dict, Optional import numpy as np import datasets _SCREAMING_SNAKE_CASE : Optional[Any] = '\nIoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union\nbetween the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation,\nthe mean IoU of the image is calculated by taking the IoU of each class and averaging them.\n' _SCREAMING_SNAKE_CASE : str = '\nArgs:\n predictions (`List[ndarray]`):\n List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n references (`List[ndarray]`):\n List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n num_labels (`int`):\n Number of classes (categories).\n ignore_index (`int`):\n Index that will be ignored during evaluation.\n nan_to_num (`int`, *optional*):\n If specified, NaN values will be replaced by the number defined by the user.\n label_map (`dict`, *optional*):\n If specified, dictionary mapping old label indices to new label indices.\n reduce_labels (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background,\n and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255.\n\nReturns:\n `Dict[str, float | ndarray]` comprising various elements:\n - *mean_iou* (`float`):\n Mean Intersection-over-Union (IoU averaged over all categories).\n - *mean_accuracy* (`float`):\n Mean accuracy (averaged over all categories).\n - *overall_accuracy* (`float`):\n Overall accuracy on all images.\n - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`):\n Per category accuracy.\n - *per_category_iou* (`ndarray` of shape `(num_labels,)`):\n Per category IoU.\n\nExamples:\n\n >>> import numpy as np\n\n >>> mean_iou = datasets.load_metric("mean_iou")\n\n >>> # suppose one has 3 different segmentation maps predicted\n >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]])\n >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]])\n\n >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]])\n >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]])\n\n >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]])\n >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]])\n\n >>> predicted = [predicted_1, predicted_2, predicted_3]\n >>> ground_truth = [actual_1, actual_2, actual_3]\n\n >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False)\n >>> print(results) # doctest: +NORMALIZE_WHITESPACE\n {\'mean_iou\': 0.47750000000000004, \'mean_accuracy\': 0.5916666666666666, \'overall_accuracy\': 0.5263157894736842, \'per_category_iou\': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), \'per_category_accuracy\': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])}\n' _SCREAMING_SNAKE_CASE : Optional[Any] = '\\n@software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020,\nauthor = {{MMSegmentation Contributors}},\nlicense = {Apache-2.0},\nmonth = {7},\ntitle = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}},\nurl = {https://github.com/open-mmlab/mmsegmentation},\nyear = {2020}\n}' def UpperCAmelCase_ ( _A , _A , _A , _A , _A = None , _A = False , ): '''simple docstring''' if label_map is not None: for old_id, new_id in label_map.items(): SCREAMING_SNAKE_CASE__ = new_id # turn into Numpy arrays SCREAMING_SNAKE_CASE__ = np.array(_A ) SCREAMING_SNAKE_CASE__ = np.array(_A ) if reduce_labels: SCREAMING_SNAKE_CASE__ = 2_55 SCREAMING_SNAKE_CASE__ = label - 1 SCREAMING_SNAKE_CASE__ = 2_55 SCREAMING_SNAKE_CASE__ = label != ignore_index SCREAMING_SNAKE_CASE__ = np.not_equal(_A , _A ) SCREAMING_SNAKE_CASE__ = pred_label[mask] SCREAMING_SNAKE_CASE__ = np.array(_A )[mask] SCREAMING_SNAKE_CASE__ = pred_label[pred_label == label] SCREAMING_SNAKE_CASE__ = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] SCREAMING_SNAKE_CASE__ = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] SCREAMING_SNAKE_CASE__ = np.histogram(_A , bins=_A , range=(0, num_labels - 1) )[0] SCREAMING_SNAKE_CASE__ = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def UpperCAmelCase_ ( _A , _A , _A , _A , _A = None , _A = False , ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = np.zeros((num_labels,) , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = np.zeros((num_labels,) , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = np.zeros((num_labels,) , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(_A , _A ): SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = intersect_and_union( _A , _A , _A , _A , _A , _A ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def UpperCAmelCase_ ( _A , _A , _A , _A , _A = None , _A = None , _A = False , ): '''simple docstring''' SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = total_intersect_and_union( _A , _A , _A , _A , _A , _A ) # compute metrics SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = total_area_intersect.sum() / total_area_label.sum() SCREAMING_SNAKE_CASE__ = total_area_intersect / total_area_union SCREAMING_SNAKE_CASE__ = total_area_intersect / total_area_label SCREAMING_SNAKE_CASE__ = np.nanmean(_A ) SCREAMING_SNAKE_CASE__ = np.nanmean(_A ) SCREAMING_SNAKE_CASE__ = all_acc SCREAMING_SNAKE_CASE__ = iou SCREAMING_SNAKE_CASE__ = acc if nan_to_num is not None: SCREAMING_SNAKE_CASE__ = {metric: np.nan_to_num(_A , nan=_A ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase__ ( datasets.Metric ): """simple docstring""" def lowercase_ ( self : Any ) -> Union[str, Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { '''predictions''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16''' ) ) ), '''references''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16''' ) ) ), } ) , reference_urls=[ '''https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py''' ] , ) def lowercase_ ( self : Optional[int] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any = None , __lowerCamelCase : Tuple = None , __lowerCamelCase : Optional[Any] = False , ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = mean_iou( results=lowercase__ , gt_seg_maps=lowercase__ , num_labels=lowercase__ , ignore_index=lowercase__ , nan_to_num=lowercase__ , label_map=lowercase__ , reduce_labels=lowercase__ , ) return iou_result
314
def __a ( ) -> list[list[int]]: '''simple docstring''' return [list(range(1_0_0_0 - i , -1_0_0_0 - i , -1 ) ) for i in range(1_0_0_0 )] A_ : Union[str, Any] = generate_large_matrix() A_ : Union[str, Any] = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def __a ( SCREAMING_SNAKE_CASE ) -> None: '''simple docstring''' assert all(row == sorted(SCREAMING_SNAKE_CASE , reverse=SCREAMING_SNAKE_CASE ) for row in grid ) assert all(list(SCREAMING_SNAKE_CASE ) == sorted(SCREAMING_SNAKE_CASE , reverse=SCREAMING_SNAKE_CASE ) for col in zip(*SCREAMING_SNAKE_CASE ) ) def __a ( SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' __UpperCAmelCase = 0 __UpperCAmelCase = len(SCREAMING_SNAKE_CASE ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: __UpperCAmelCase = (left + right) // 2 __UpperCAmelCase = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: __UpperCAmelCase = mid + 1 else: __UpperCAmelCase = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(SCREAMING_SNAKE_CASE ) def __a ( SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' __UpperCAmelCase = 0 __UpperCAmelCase = len(grid[0] ) for i in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = find_negative_index(grid[i][:bound] ) total += bound return (len(SCREAMING_SNAKE_CASE ) * len(grid[0] )) - total def __a ( SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' return len([number for row in grid for number in row if number < 0] ) def __a ( SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' __UpperCAmelCase = 0 for row in grid: for i, number in enumerate(SCREAMING_SNAKE_CASE ): if number < 0: total += len(SCREAMING_SNAKE_CASE ) - i break return total def __a ( ) -> None: '''simple docstring''' from timeit import timeit print('''Running benchmarks''' ) __UpperCAmelCase = ( '''from __main__ import count_negatives_binary_search, ''' '''count_negatives_brute_force, count_negatives_brute_force_with_break, grid''' ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): __UpperCAmelCase = timeit(f'''{func}(grid=grid)''' , setup=SCREAMING_SNAKE_CASE , number=5_0_0 ) print(f'''{func}() took {time:0.4f} seconds''' ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
333
0
"""simple docstring""" import numpy as np from transformers import Pipeline def snake_case_ ( lowerCAmelCase_ : Tuple ): __lowercase : int = np.max(a__ , axis=-1 , keepdims=a__ ) __lowercase : List[Any] = np.exp(outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=a__ ) class lowerCAmelCase ( _a ): '''simple docstring''' def lowerCAmelCase ( self : List[str] , **__a : Tuple ) -> Union[str, Any]: """simple docstring""" __lowercase : Optional[int] = {} if "second_text" in kwargs: __lowercase : Optional[int] = kwargs["""second_text"""] return preprocess_kwargs, {}, {} def lowerCAmelCase ( self : List[str] , __a : Dict , __a : Any=None ) -> List[str]: """simple docstring""" return self.tokenizer(__lowerCAmelCase , text_pair=__lowerCAmelCase , return_tensors=self.framework ) def lowerCAmelCase ( self : Optional[int] , __a : Union[str, Any] ) -> Tuple: """simple docstring""" return self.model(**__lowerCAmelCase ) def lowerCAmelCase ( self : str , __a : Union[str, Any] ) -> Dict: """simple docstring""" __lowercase : Optional[Any] = model_outputs.logits[0].numpy() __lowercase : Any = softmax(__lowerCAmelCase ) __lowercase : Dict = np.argmax(__lowerCAmelCase ) __lowercase : int = self.model.config.idalabel[best_class] __lowercase : Any = probabilities[best_class].item() __lowercase : str = logits.tolist() return {"label": label, "score": score, "logits": logits}
371
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase : str = { '''facebook/nllb-moe-54B''': '''https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json''', } class lowerCAmelCase ( __a ): '''simple docstring''' _A : int = '''nllb-moe''' _A : List[str] = ['''past_key_values'''] _A : Optional[Any] = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Dict , __a : List[str]=128112 , __a : List[Any]=1024 , __a : List[Any]=12 , __a : Union[str, Any]=4096 , __a : List[str]=16 , __a : int=12 , __a : Optional[int]=4096 , __a : str=16 , __a : List[Any]=0.05 , __a : Any=0.05 , __a : Dict=True , __a : Optional[Any]=True , __a : List[Any]="relu" , __a : Tuple=1024 , __a : Optional[Any]=0.1 , __a : Tuple=0.1 , __a : Any=0.0 , __a : Optional[Any]=0.02 , __a : List[str]=2 , __a : Union[str, Any]=True , __a : List[Any]=False , __a : Tuple="float32" , __a : Optional[int]=False , __a : Optional[int]=128 , __a : str=64 , __a : Dict=4 , __a : str=4 , __a : List[str]=0.001 , __a : List[Any]=0.001 , __a : Optional[Any]="all" , __a : Optional[int]=False , __a : int=False , __a : int=1.0 , __a : Dict=0.2 , __a : Tuple=1 , __a : Optional[Any]=0 , __a : List[Any]=2 , __a : Any=False , **__a : Any , ) -> Any: """simple docstring""" __lowercase : int = vocab_size __lowercase : List[Any] = max_position_embeddings __lowercase : Tuple = d_model __lowercase : str = encoder_ffn_dim __lowercase : List[str] = encoder_layers __lowercase : int = encoder_attention_heads __lowercase : List[Any] = decoder_ffn_dim __lowercase : int = decoder_layers __lowercase : Optional[int] = decoder_attention_heads __lowercase : Union[str, Any] = dropout __lowercase : str = attention_dropout __lowercase : Any = activation_dropout __lowercase : List[Any] = activation_function __lowercase : List[str] = init_std __lowercase : Optional[int] = encoder_layerdrop __lowercase : str = decoder_layerdrop __lowercase : Dict = use_cache __lowercase : Optional[Any] = encoder_layers __lowercase : str = scale_embedding # scale factor will be sqrt(d_model) if True __lowercase : List[Any] = router_z_loss_coef __lowercase : Tuple = router_aux_loss_coef __lowercase : str = decoder_sparse_step __lowercase : Any = encoder_sparse_step __lowercase : str = num_experts __lowercase : List[Any] = expert_capacity __lowercase : int = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) __lowercase : Optional[int] = router_dtype __lowercase : Any = router_ignore_padding_tokens __lowercase : Optional[Any] = batch_prioritized_routing __lowercase : str = second_expert_policy __lowercase : List[str] = normalize_router_prob_before_dropping __lowercase : List[Any] = moe_eval_capacity_token_fraction __lowercase : List[str] = moe_token_dropout __lowercase : Optional[Any] = output_router_logits super().__init__( pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , is_encoder_decoder=__a , decoder_start_token_id=__a , **__a , )
306
0
'''simple docstring''' import datasets from .evaluate import evaluate _A : Tuple ='''\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } ''' _A : int =''' This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. ''' _A : Dict =''' Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair as given in the references (see below) - \'prediction_text\': the text of the answer references: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair (see above), - \'answers\': a Dict in the SQuAD dataset format { \'text\': list of possible texts for the answer, as a list of strings \'answer_start\': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: \'exact_match\': Exact match (the normalized answer exactly match the gold answer) \'f1\': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}] >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}] >>> squad_metric = datasets.load_metric("squad") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 100.0, \'f1\': 100.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def lowerCamelCase_ ( self: List[str] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Tuple ): lowerCamelCase__ : List[str] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} lowerCamelCase__ : Tuple = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] lowerCamelCase__ : Any = evaluate(dataset=UpperCamelCase__ , predictions=UpperCamelCase__ ) return score
41
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { 'BridgeTower/bridgetower-base': 'https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json', 'BridgeTower/bridgetower-base-itm-mlm': ( 'https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json' ), } class A__ ( UpperCamelCase ): """simple docstring""" UpperCamelCase_ : Optional[Any] = '''bridgetower_vision_model''' def __init__( self : int , lowerCAmelCase__ : List[Any]=7_6_8 , lowerCAmelCase__ : Dict=1_2 , lowerCAmelCase__ : Union[str, Any]=3 , lowerCAmelCase__ : Dict=1_6 , lowerCAmelCase__ : int=2_8_8 , lowerCAmelCase__ : Dict=1 , lowerCAmelCase__ : int=1e-05 , lowerCAmelCase__ : Optional[Any]=False , lowerCAmelCase__ : int=True , lowerCAmelCase__ : int=False , **lowerCAmelCase__ : Union[str, Any] , ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowerCAmelCase__ ) _UpperCAmelCase : Optional[int] = hidden_size _UpperCAmelCase : Dict = num_hidden_layers _UpperCAmelCase : Dict = num_channels _UpperCAmelCase : Tuple = patch_size _UpperCAmelCase : str = image_size _UpperCAmelCase : List[Any] = initializer_factor _UpperCAmelCase : Any = layer_norm_eps _UpperCAmelCase : Optional[Any] = stop_gradient _UpperCAmelCase : List[str] = share_layernorm _UpperCAmelCase : List[str] = remove_last_layer @classmethod def _lowerCAmelCase ( cls : Optional[Any] , lowerCAmelCase__ : Union[str, os.PathLike] , **lowerCAmelCase__ : Any ) -> "PretrainedConfig": """simple docstring""" _UpperCAmelCase , _UpperCAmelCase : Any = cls.get_config_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) if config_dict.get("model_type" ) == "bridgetower": _UpperCAmelCase : Optional[Any] = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) class A__ ( UpperCamelCase ): """simple docstring""" UpperCamelCase_ : Optional[int] = '''bridgetower_text_model''' def __init__( self : int , lowerCAmelCase__ : Optional[int]=5_0_2_6_5 , lowerCAmelCase__ : Tuple=7_6_8 , lowerCAmelCase__ : List[str]=1_2 , lowerCAmelCase__ : int=1_2 , lowerCAmelCase__ : str=1 , lowerCAmelCase__ : Optional[Any]=3_0_7_2 , lowerCAmelCase__ : List[Any]="gelu" , lowerCAmelCase__ : str=0.1 , lowerCAmelCase__ : Optional[Any]=0.1 , lowerCAmelCase__ : int=5_1_4 , lowerCAmelCase__ : List[Any]=1 , lowerCAmelCase__ : Any=1e-05 , lowerCAmelCase__ : Optional[Any]=1 , lowerCAmelCase__ : List[Any]=0 , lowerCAmelCase__ : Dict=2 , lowerCAmelCase__ : List[Any]="absolute" , lowerCAmelCase__ : Optional[Any]=True , **lowerCAmelCase__ : Any , ) -> List[Any]: """simple docstring""" super().__init__(**lowerCAmelCase__ ) _UpperCAmelCase : List[str] = vocab_size _UpperCAmelCase : Optional[Any] = hidden_size _UpperCAmelCase : Tuple = num_hidden_layers _UpperCAmelCase : Dict = num_attention_heads _UpperCAmelCase : Dict = hidden_act _UpperCAmelCase : int = initializer_factor _UpperCAmelCase : List[str] = intermediate_size _UpperCAmelCase : Optional[Any] = hidden_dropout_prob _UpperCAmelCase : Any = attention_probs_dropout_prob _UpperCAmelCase : Tuple = max_position_embeddings _UpperCAmelCase : Any = type_vocab_size _UpperCAmelCase : Union[str, Any] = layer_norm_eps _UpperCAmelCase : Optional[Any] = position_embedding_type _UpperCAmelCase : Optional[int] = use_cache _UpperCAmelCase : Optional[Any] = pad_token_id _UpperCAmelCase : Union[str, Any] = bos_token_id _UpperCAmelCase : int = eos_token_id @classmethod def _lowerCAmelCase ( cls : Tuple , lowerCAmelCase__ : Union[str, os.PathLike] , **lowerCAmelCase__ : Dict ) -> "PretrainedConfig": """simple docstring""" _UpperCAmelCase , _UpperCAmelCase : str = cls.get_config_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) if config_dict.get("model_type" ) == "bridgetower": _UpperCAmelCase : int = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCAmelCase__ , **lowerCAmelCase__ ) class A__ ( UpperCamelCase ): """simple docstring""" UpperCamelCase_ : Any = '''bridgetower''' def __init__( self : List[str] , lowerCAmelCase__ : int=True , lowerCAmelCase__ : Any="gelu" , lowerCAmelCase__ : Dict=7_6_8 , lowerCAmelCase__ : Any=1 , lowerCAmelCase__ : List[str]=1e-05 , lowerCAmelCase__ : int=False , lowerCAmelCase__ : str="add" , lowerCAmelCase__ : Tuple=1_2 , lowerCAmelCase__ : Optional[int]=6 , lowerCAmelCase__ : Optional[Any]=False , lowerCAmelCase__ : Union[str, Any]=False , lowerCAmelCase__ : Tuple=None , lowerCAmelCase__ : List[Any]=None , **lowerCAmelCase__ : Optional[Any] , ) -> Tuple: """simple docstring""" _UpperCAmelCase : Optional[int] = kwargs.pop("text_config_dict" , lowerCAmelCase__ ) _UpperCAmelCase : int = kwargs.pop("vision_config_dict" , lowerCAmelCase__ ) super().__init__(**lowerCAmelCase__ ) _UpperCAmelCase : Optional[int] = share_cross_modal_transformer_layers _UpperCAmelCase : int = hidden_act _UpperCAmelCase : str = hidden_size _UpperCAmelCase : Tuple = initializer_factor _UpperCAmelCase : Union[str, Any] = layer_norm_eps _UpperCAmelCase : Tuple = share_link_tower_layers _UpperCAmelCase : List[str] = link_tower_type _UpperCAmelCase : int = num_attention_heads _UpperCAmelCase : Any = num_hidden_layers _UpperCAmelCase : Optional[int] = tie_word_embeddings _UpperCAmelCase : int = init_layernorm_from_vision_encoder if text_config is None: _UpperCAmelCase : str = {} logger.info("`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values." ) if vision_config is None: _UpperCAmelCase : Union[str, Any] = {} logger.info("`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values." ) _UpperCAmelCase : str = BridgeTowerTextConfig(**lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = BridgeTowerVisionConfig(**lowerCAmelCase__ ) @classmethod def _lowerCAmelCase ( cls : Union[str, Any] , lowerCAmelCase__ : BridgeTowerTextConfig , lowerCAmelCase__ : BridgeTowerVisionConfig , **lowerCAmelCase__ : Union[str, Any] ) -> Any: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowerCAmelCase__ ) def _lowerCAmelCase ( self : str ) -> Tuple: """simple docstring""" _UpperCAmelCase : str = copy.deepcopy(self.__dict__ ) _UpperCAmelCase : Union[str, Any] = self.text_config.to_dict() _UpperCAmelCase : Union[str, Any] = self.vision_config.to_dict() _UpperCAmelCase : List[str] = self.__class__.model_type return output
145
0
class a__ : def __init__( self , UpperCAmelCase ) -> None: __a = size __a = [0] * size __a = [0] * size @staticmethod def __SCREAMING_SNAKE_CASE ( UpperCAmelCase ) -> int: return index | (index + 1) @staticmethod def __SCREAMING_SNAKE_CASE ( UpperCAmelCase ) -> int: return (index & (index + 1)) - 1 def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase ) -> None: __a = value while index < self.size: __a = self.get_prev(UpperCAmelCase ) + 1 if current_left_border == index: __a = value else: __a = max(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) __a = self.get_next(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase ) -> int: right -= 1 # Because of right is exclusive __a = 0 while left <= right: __a = self.get_prev(UpperCAmelCase ) if left <= current_left: __a = max(UpperCAmelCase , self.tree[right] ) __a = current_left else: __a = max(UpperCAmelCase , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
197
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase_ : Dict = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Any = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
197
1
import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( '''split_dict''' , [ SplitDict(), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1_337 , num_examples=42 , dataset_name='''my_dataset''' )} ), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1_337 , num_examples=42 )} ), SplitDict({'''train''': SplitInfo()} ), ] , ) def A ( _lowercase ): SCREAMING_SNAKE_CASE : Tuple = split_dict._to_yaml_list() assert len(_lowercase ) == len(_lowercase ) SCREAMING_SNAKE_CASE : Tuple = SplitDict._from_yaml_list(_lowercase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump SCREAMING_SNAKE_CASE : Any = None # the split name of split_dict takes over the name of the split info object SCREAMING_SNAKE_CASE : Optional[Any] = split_name assert split_dict == reloaded @pytest.mark.parametrize( '''split_info''' , [SplitInfo(), SplitInfo(dataset_name=_lowercase ), SplitInfo(dataset_name='''my_dataset''' )] ) def A ( _lowercase ): # For backward compatibility, we need asdict(split_dict) to return split info dictrionaries with the "dataset_name" # field even if it's deprecated. This way old versionso of `datasets` can still reload dataset_infos.json files SCREAMING_SNAKE_CASE : List[Any] = asdict(SplitDict({'''train''': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
182
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCamelCase : List[Any] = {'configuration_glpn': ['GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GLPNConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Optional[Any] = ['GLPNFeatureExtractor'] __UpperCamelCase : List[str] = ['GLPNImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : str = [ 'GLPN_PRETRAINED_MODEL_ARCHIVE_LIST', 'GLPNForDepthEstimation', 'GLPNLayer', 'GLPNModel', 'GLPNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys __UpperCamelCase : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
182
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase : Optional[Any] = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : List[str] = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys UpperCamelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
357
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( snake_case : int , snake_case : int ) -> bool: """simple docstring""" return numa ^ numa < 0 if __name__ == "__main__": import doctest doctest.testmod()
345
0
'''simple docstring''' import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): a_ : List[str] = yaml.safe_load( """\ name: \"\" allow_empty: false allow_empty_text: true subsections: - name: \"Dataset Card for X\" # First-level markdown heading allow_empty: false allow_empty_text: true subsections: - name: \"Table of Contents\" allow_empty: false allow_empty_text: false subsections: null - name: \"Dataset Description\" allow_empty: false allow_empty_text: false subsections: - name: \"Dataset Summary\" allow_empty: false allow_empty_text: false subsections: null - name: \"Supported Tasks and Leaderboards\" allow_empty: true allow_empty_text: true subsections: null - name: Languages allow_empty: false allow_empty_text: true subsections: null """ ) a_ : int = { """name""": """root""", """text""": """""", """is_empty_text""": True, """subsections""": [ { """name""": """Dataset Card for My Dataset""", """text""": """""", """is_empty_text""": True, """subsections""": [ {"""name""": """Table of Contents""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": []}, { """name""": """Dataset Description""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [ { """name""": """Dataset Summary""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [], }, { """name""": """Supported Tasks and Leaderboards""", """text""": """""", """is_empty_text""": True, """subsections""": [], }, {"""name""": """Languages""", """text""": """Language Text""", """is_empty_text""": False, """subsections""": []}, ], }, ], } ], } a_ : List[str] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : str = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. #### Extra Ignored Subsection ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : List[str] = { """name""": """root""", """text""": """""", """is_empty_text""": True, """subsections""": [ { """name""": """Dataset Card for My Dataset""", """text""": """""", """is_empty_text""": True, """subsections""": [ {"""name""": """Table of Contents""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": []}, { """name""": """Dataset Description""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [ { """name""": """Dataset Summary""", """text""": """Some text here.""", """is_empty_text""": False, """subsections""": [ { """name""": """Extra Ignored Subsection""", """text""": """""", """is_empty_text""": True, """subsections""": [], } ], }, { """name""": """Supported Tasks and Leaderboards""", """text""": """""", """is_empty_text""": True, """subsections""": [], }, {"""name""": """Languages""", """text""": """Language Text""", """is_empty_text""": False, """subsections""": []}, ], }, ], } ], } a_ : str = """\ --- --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : Tuple = ( """The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README.""" ) a_ : List[Any] = """\ # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : Dict = ( """The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README.""" ) a_ : List[str] = """\ --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : str = """The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README.""" a_ : Optional[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : int = """The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored).""" a_ : Optional[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset """ a_ : List[Any] = """The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found 'None'.""" a_ : Any = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Languages Language Text """ a_ : Optional[Any] = """The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`.""" a_ : int = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages """ a_ : str = """The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty.""" a_ : int = """\ --- language: - zh - en --- ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : List[Any] = """The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.""" a_ : int = """\ --- language: - zh - en --- # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text # Dataset Card My Dataset """ a_ : Optional[Any] = """The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README.""" a_ : Dict = """\ --- language: - zh - en --- # Dataset Card My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : List[Any] = """The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README.""" a_ : Optional[Any] = """""" a_ : Tuple = """The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README.""" a_ : Optional[Any] = """\ --- language: - zh - en --- # Dataset Card for My Dataset # Dataset Card for My Dataset ## Table of Contents Some text here. ## Dataset Description Some text here. ### Dataset Summary Some text here. ### Supported Tasks and Leaderboards ### Languages Language Text """ a_ : Dict = """The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections.""" @pytest.mark.parametrize( '''readme_md, expected_dict''' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def a_ ( __snake_case : Tuple , __snake_case : str ) -> str: """simple docstring""" assert ReadMe.from_string(__snake_case , __snake_case ).to_dict() == expected_dict @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def a_ ( __snake_case : List[str] , __snake_case : Union[str, Any] ) -> Optional[int]: """simple docstring""" with pytest.raises(__snake_case , match=re.escape(expected_error.format(path='''root''' ) ) ): lowerCamelCase_ =ReadMe.from_string(__snake_case , __snake_case ) readme.validate() @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def a_ ( __snake_case : Union[str, Any] , __snake_case : Dict ) -> List[str]: """simple docstring""" with pytest.raises(__snake_case , match=re.escape(expected_error.format(path='''root''' ) ) ): ReadMe.from_string(__snake_case , __snake_case ) @pytest.mark.parametrize( '''readme_md,''' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def a_ ( __snake_case : Tuple ) -> int: """simple docstring""" ReadMe.from_string(__snake_case , __snake_case , suppress_parsing_errors=__snake_case ) @pytest.mark.parametrize( '''readme_md, expected_dict''' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def a_ ( __snake_case : Tuple , __snake_case : Tuple ) -> Union[str, Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase_ =Path(__snake_case ) / '''README.md''' with open(__snake_case , '''w+''' ) as readme_file: readme_file.write(__snake_case ) lowerCamelCase_ =ReadMe.from_readme(__snake_case , __snake_case ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def a_ ( __snake_case : Union[str, Any] , __snake_case : int ) -> str: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase_ =Path(__snake_case ) / '''README.md''' with open(__snake_case , '''w+''' ) as readme_file: readme_file.write(__snake_case ) lowerCamelCase_ =expected_error.format(path=__snake_case ) with pytest.raises(__snake_case , match=re.escape(__snake_case ) ): lowerCamelCase_ =ReadMe.from_readme(__snake_case , __snake_case ) readme.validate() @pytest.mark.parametrize( '''readme_md, expected_error''' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def a_ ( __snake_case : List[Any] , __snake_case : Union[str, Any] ) -> List[Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase_ =Path(__snake_case ) / '''README.md''' with open(__snake_case , '''w+''' ) as readme_file: readme_file.write(__snake_case ) lowerCamelCase_ =expected_error.format(path=__snake_case ) with pytest.raises(__snake_case , match=re.escape(__snake_case ) ): ReadMe.from_readme(__snake_case , __snake_case ) @pytest.mark.parametrize( '''readme_md,''' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def a_ ( __snake_case : List[Any] ) -> List[str]: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase_ =Path(__snake_case ) / '''README.md''' with open(__snake_case , '''w+''' ) as readme_file: readme_file.write(__snake_case ) ReadMe.from_readme(__snake_case , __snake_case , suppress_parsing_errors=__snake_case )
75
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'kssteven/ibert-roberta-base': 'https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json', 'kssteven/ibert-roberta-large': 'https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json', 'kssteven/ibert-roberta-large-mnli': ( 'https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json' ), } class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): snake_case_ = """ibert""" def __init__( self : Optional[int] , __lowercase : List[str]=3_05_22 , __lowercase : Tuple=7_68 , __lowercase : str=12 , __lowercase : Optional[int]=12 , __lowercase : Optional[Any]=30_72 , __lowercase : str="gelu" , __lowercase : List[str]=0.1 , __lowercase : List[Any]=0.1 , __lowercase : List[str]=5_12 , __lowercase : str=2 , __lowercase : Tuple=0.02 , __lowercase : Union[str, Any]=1e-12 , __lowercase : List[Any]=1 , __lowercase : List[str]=0 , __lowercase : Optional[Any]=2 , __lowercase : int="absolute" , __lowercase : Tuple=False , __lowercase : int="none" , **__lowercase : Optional[Any] , ) -> List[Any]: super().__init__(pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase , **__lowercase ) SCREAMING_SNAKE_CASE__ : Any =vocab_size SCREAMING_SNAKE_CASE__ : Dict =hidden_size SCREAMING_SNAKE_CASE__ : str =num_hidden_layers SCREAMING_SNAKE_CASE__ : Union[str, Any] =num_attention_heads SCREAMING_SNAKE_CASE__ : Tuple =hidden_act SCREAMING_SNAKE_CASE__ : List[str] =intermediate_size SCREAMING_SNAKE_CASE__ : Union[str, Any] =hidden_dropout_prob SCREAMING_SNAKE_CASE__ : int =attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : Union[str, Any] =max_position_embeddings SCREAMING_SNAKE_CASE__ : Dict =type_vocab_size SCREAMING_SNAKE_CASE__ : Tuple =initializer_range SCREAMING_SNAKE_CASE__ : str =layer_norm_eps SCREAMING_SNAKE_CASE__ : Tuple =position_embedding_type SCREAMING_SNAKE_CASE__ : Any =quant_mode SCREAMING_SNAKE_CASE__ : Optional[int] =force_dequant class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): @property def __magic_name__ ( self : str ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": SCREAMING_SNAKE_CASE__ : str ={0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: SCREAMING_SNAKE_CASE__ : Any ={0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
152
0
"""simple docstring""" from typing import Dict from .base import GenericTensor, Pipeline class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self , __A=None , __A=None , __A=None , **__A ) -> Tuple: if tokenize_kwargs is None: a ={} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( '''truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)''' ) a =truncation a =tokenize_kwargs a ={} if return_tensors is not None: a =return_tensors return preprocess_params, {}, postprocess_params def SCREAMING_SNAKE_CASE ( self , __A , **__A ) -> Dict[str, GenericTensor]: a =self.framework a =self.tokenizer(__A , return_tensors=__A , **__A ) return model_inputs def SCREAMING_SNAKE_CASE ( self , __A ) -> Optional[Any]: a =self.model(**__A ) return model_outputs def SCREAMING_SNAKE_CASE ( self , __A , __A=False ) -> Tuple: # [0] is the first available tensor, logits or last_hidden_state. if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self , *__A , **__A ) -> Dict: return super().__call__(*__A , **__A )
215
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCamelCase_ : List[Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = ["pixel_values"] def __init__( self , __A = True , __A = None , __A = PILImageResampling.BICUBIC , __A = True , __A = None , __A = True , __A = 1 / 255 , __A = True , __A = None , __A = None , __A = True , **__A , ) -> None: super().__init__(**__A ) a =size if size is not None else {'''shortest_edge''': 224} a =get_size_dict(__A , default_to_square=__A ) a =crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} a =get_size_dict(__A , default_to_square=__A , param_name='''crop_size''' ) a =do_resize a =size a =resample a =do_center_crop a =crop_size a =do_rescale a =rescale_factor a =do_normalize a =image_mean if image_mean is not None else OPENAI_CLIP_MEAN a =image_std if image_std is not None else OPENAI_CLIP_STD a =do_convert_rgb def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = PILImageResampling.BICUBIC , __A = None , **__A , ) -> np.ndarray: a =get_size_dict(__A , default_to_square=__A ) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) a =get_resize_output_image_size(__A , size=size['''shortest_edge'''] , default_to_square=__A ) return resize(__A , size=__A , resample=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None , **__A , ) -> np.ndarray: a =get_size_dict(__A ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(__A , size=(size['''height'''], size['''width''']) , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None , **__A , ) -> Any: return rescale(__A , scale=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A = None , **__A , ) -> np.ndarray: return normalize(__A , mean=__A , std=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = ChannelDimension.FIRST , **__A , ) -> PIL.Image.Image: a =do_resize if do_resize is not None else self.do_resize a =size if size is not None else self.size a =get_size_dict(__A , param_name='''size''' , default_to_square=__A ) a =resample if resample is not None else self.resample a =do_center_crop if do_center_crop is not None else self.do_center_crop a =crop_size if crop_size is not None else self.crop_size a =get_size_dict(__A , param_name='''crop_size''' , default_to_square=__A ) a =do_rescale if do_rescale is not None else self.do_rescale a =rescale_factor if rescale_factor is not None else self.rescale_factor a =do_normalize if do_normalize is not None else self.do_normalize a =image_mean if image_mean is not None else self.image_mean a =image_std if image_std is not None else self.image_std a =do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb a =make_list_of_images(__A ) if not valid_images(__A ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: a =[convert_to_rgb(__A ) for image in images] # All transformations expect numpy arrays. a =[to_numpy_array(__A ) for image in images] if do_resize: a =[self.resize(image=__A , size=__A , resample=__A ) for image in images] if do_center_crop: a =[self.center_crop(image=__A , size=__A ) for image in images] if do_rescale: a =[self.rescale(image=__A , scale=__A ) for image in images] if do_normalize: a =[self.normalize(image=__A , mean=__A , std=__A ) for image in images] a =[to_channel_dimension_format(__A , __A ) for image in images] a ={'''pixel_values''': images} return BatchFeature(data=__A , tensor_type=__A )
215
1
from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def _A ( SCREAMING_SNAKE_CASE : Namespace ): """simple docstring""" return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) UpperCAmelCase : Optional[int] = """ transformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires TensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions. """ class __lowerCAmelCase ( UpperCamelCase__): @staticmethod def _lowercase ( lowerCAmelCase__ ) -> Any: '''simple docstring''' a__ : int =parser.add_parser( "convert" , help="CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints." , ) train_parser.add_argument("--model_type" , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help="Model's type." ) train_parser.add_argument( "--tf_checkpoint" , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help="TensorFlow checkpoint path or folder." ) train_parser.add_argument( "--pytorch_dump_output" , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help="Path to the PyTorch saved model output." ) train_parser.add_argument("--config" , type=lowerCAmelCase__ , default="" , help="Configuration file path or folder." ) train_parser.add_argument( "--finetuning_task_name" , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help="Optional fine-tuning task name if the TF model was a finetuned model." , ) train_parser.set_defaults(func=lowerCAmelCase__ ) def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , *lowerCAmelCase__ , ) -> Optional[int]: '''simple docstring''' a__ : Optional[int] =logging.get_logger("transformers-cli/converting" ) self._logger.info(F'''Loading model {model_type}''' ) a__ : Optional[int] =model_type a__ : Union[str, Any] =tf_checkpoint a__ : Optional[Any] =pytorch_dump_output a__ : Dict =config a__ : Tuple =finetuning_task_name def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(lowerCAmelCase__ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) if "ckpt" in self._tf_checkpoint.lower(): a__ : Tuple =self._tf_checkpoint a__ : str ="" else: a__ : Union[str, Any] =self._tf_checkpoint a__ : List[Any] ="" convert_transfo_xl_checkpoint_to_pytorch( lowerCAmelCase__ , self._config , self._pytorch_dump_output , lowerCAmelCase__ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(lowerCAmelCase__ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( "--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]" )
95
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[int] = { """google/canine-s""": """https://huggingface.co/google/canine-s/resolve/main/config.json""", # See all CANINE models at https://huggingface.co/models?filter=canine } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : List[Any] = """canine""" def __init__( self , lowerCAmelCase__=7_6_8 , lowerCAmelCase__=1_2 , lowerCAmelCase__=1_2 , lowerCAmelCase__=3_0_7_2 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1_6_3_8_4 , lowerCAmelCase__=1_6 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0 , lowerCAmelCase__=0XE0_00 , lowerCAmelCase__=0XE0_01 , lowerCAmelCase__=4 , lowerCAmelCase__=4 , lowerCAmelCase__=8 , lowerCAmelCase__=1_6_3_8_4 , lowerCAmelCase__=1_2_8 , **lowerCAmelCase__ , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a__ : Optional[int] =max_position_embeddings a__ : str =hidden_size a__ : Optional[Any] =num_hidden_layers a__ : Tuple =num_attention_heads a__ : Optional[Any] =intermediate_size a__ : Optional[int] =hidden_act a__ : List[Any] =hidden_dropout_prob a__ : Union[str, Any] =attention_probs_dropout_prob a__ : Optional[Any] =initializer_range a__ : Union[str, Any] =type_vocab_size a__ : Optional[int] =layer_norm_eps # Character config: a__ : int =downsampling_rate a__ : Optional[Any] =upsampling_kernel_size a__ : Union[str, Any] =num_hash_functions a__ : Any =num_hash_buckets a__ : int =local_transformer_stride
95
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = """▁""" _SCREAMING_SNAKE_CASE = {"""vocab_file""": """sentencepiece.bpe.model"""} _SCREAMING_SNAKE_CASE = { """vocab_file""": { """facebook/xglm-564M""": """https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model""", } } _SCREAMING_SNAKE_CASE = { """facebook/xglm-564M""": 2_0_4_8, } class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any] , lowerCamelCase_ : Any , lowerCamelCase_ : Optional[Any]="<s>" , lowerCamelCase_ : Dict="</s>" , lowerCamelCase_ : str="</s>" , lowerCamelCase_ : Dict="<s>" , lowerCamelCase_ : List[str]="<unk>" , lowerCamelCase_ : Optional[int]="<pad>" , lowerCamelCase_ : Optional[Dict[str, Any]] = None , **lowerCamelCase_ : str , ): """simple docstring""" UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer UpperCamelCase = 7 UpperCamelCase = [f"""<madeupword{i}>""" for i in range(self.num_madeup_words )] UpperCamelCase = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase_ ) ) UpperCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCamelCase = 1 # Mimic fairseq token-to-id alignment for the first 4 token UpperCamelCase = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} UpperCamelCase = len(self.sp_model ) UpperCamelCase = {f"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(lowerCamelCase_ ) UpperCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : List[str] ): """simple docstring""" UpperCamelCase = self.__dict__.copy() UpperCamelCase = None UpperCamelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self : Any , lowerCamelCase_ : List[Any] ): """simple docstring""" UpperCamelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCamelCase = {} UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.sep_token_id] + token_ids_a UpperCamelCase = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def lowerCamelCase_ ( self : Dict , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None , lowerCamelCase_ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase_ )) return [1] + ([0] * len(lowerCamelCase_ )) + [1, 1] + ([0] * len(lowerCamelCase_ )) def lowerCamelCase_ ( self : Any , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None ): """simple docstring""" UpperCamelCase = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def lowerCamelCase_ ( self : List[str] ): """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = {self.convert_ids_to_tokens(lowerCamelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase_ ( self : Any , lowerCamelCase_ : str ): """simple docstring""" return self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_ ) def lowerCamelCase_ ( self : Tuple , lowerCamelCase_ : Optional[Any] ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCamelCase = self.sp_model.PieceToId(lowerCamelCase_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowerCamelCase_ ( self : str , lowerCamelCase_ : Optional[Any] ): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : Any ): """simple docstring""" UpperCamelCase = """""".join(lowerCamelCase_ ).replace(lowerCamelCase_ , """ """ ).strip() return out_string def lowerCamelCase_ ( self : Dict , lowerCamelCase_ : str , lowerCamelCase_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase_ , """wb""" ) as fi: UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_ ) return (out_vocab_file,)
350
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def lowercase( UpperCamelCase_ = True , *UpperCamelCase_ , **UpperCamelCase_ ) -> int: '''simple docstring''' if not is_tqdm_available(): raise ImportError("""Accelerate's `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`.""" ) UpperCamelCase = False if main_process_only: UpperCamelCase = PartialState().local_process_index == 0 return _tqdm(*UpperCamelCase_ , **UpperCamelCase_ , disable=UpperCamelCase_ )
165
0
from __future__ import annotations import os from collections.abc import Mapping lowerCAmelCase_ = tuple[int, int] class snake_case_ : '''simple docstring''' def __init__( self : Dict , _UpperCamelCase : set[int] , _UpperCamelCase : Mapping[EdgeT, int] ) ->None: snake_case_ = vertices snake_case_ = { (min(_UpperCamelCase ), max(_UpperCamelCase )): weight for edge, weight in edges.items() } def snake_case__( self : Union[str, Any] , _UpperCamelCase : EdgeT , _UpperCamelCase : int ) ->None: self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) snake_case_ = weight def snake_case__( self : Dict ) ->Graph: snake_case_ = Graph({min(self.vertices )} , {} ) snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 while len(subgraph.vertices ) < len(self.vertices ): snake_case_ = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: snake_case_ = edge snake_case_ = weight subgraph.add_edge(_UpperCamelCase , _UpperCamelCase ) return subgraph def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ = "p107_network.txt" ): snake_case_ = os.path.abspath(os.path.dirname(SCREAMING_SNAKE_CASE__ ) ) snake_case_ = os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) snake_case_ = {} snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 with open(SCREAMING_SNAKE_CASE__ ) as f: snake_case_ = f.read().strip().split('''\n''' ) snake_case_ = [line.split(''',''' ) for line in data] for edgea in range(1 , len(SCREAMING_SNAKE_CASE__ ) ): for edgea in range(SCREAMING_SNAKE_CASE__ ): if adjaceny_matrix[edgea][edgea] != "-": snake_case_ = int(adjaceny_matrix[edgea][edgea] ) snake_case_ = Graph(set(range(len(SCREAMING_SNAKE_CASE__ ) ) ) , SCREAMING_SNAKE_CASE__ ) snake_case_ = graph.prims_algorithm() snake_case_ = sum(graph.edges.values() ) snake_case_ = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(f"""{solution() = }""")
8
from collections import deque from .hash_table import HashTable class snake_case_ ( __A ): '''simple docstring''' def __init__( self : int , *_UpperCamelCase : int , **_UpperCamelCase : Tuple ) ->Tuple: super().__init__(*_UpperCamelCase , **_UpperCamelCase ) def snake_case__( self : Dict , _UpperCamelCase : List[str] , _UpperCamelCase : Dict ) ->Tuple: snake_case_ = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_UpperCamelCase ) snake_case_ = self.values[key] def snake_case__( self : List[Any] ) ->str: return ( sum(self.charge_factor - len(_UpperCamelCase ) for slot in self.values ) / self.size_table * self.charge_factor ) def snake_case__( self : Dict , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[int]=None ) ->str: if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_UpperCamelCase ) == 0 ): return key return super()._collision_resolution(_UpperCamelCase , _UpperCamelCase )
8
1
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger __a :str = get_logger(__name__) __a :Optional[Any] = R'\n Args:\n input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam\n search or log softmax for each vocabulary token when using beam search\n kwargs (`Dict[str, Any]`, *optional*):\n Additional logits processor specific kwargs.\n\n Return:\n `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores.\n\n' class _a : """simple docstring""" @add_start_docstrings(UpperCAmelCase ) def __call__( self : Any , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class _a : """simple docstring""" @add_start_docstrings(UpperCAmelCase ) def __call__( self : Any , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray ): raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class _a ( snake_case_ ): """simple docstring""" @add_start_docstrings(UpperCAmelCase ) def __call__( self : List[str] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int , **UpperCAmelCase : Optional[int] ): for processor in self: A_ = inspect.signature(processor.__call__ ).parameters if len(UpperCAmelCase ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( f'''Make sure that all the required parameters: {list(function_args.keys() )} for ''' f'''{processor.__class__} are passed to the logits processor.''' ) A_ = processor(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) else: A_ = processor(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase : float ): if not isinstance(UpperCAmelCase , UpperCAmelCase ) or not (temperature > 0): raise ValueError(f'''`temperature` has to be a strictly positive float, but is {temperature}''' ) A_ = temperature def __call__( self : List[Any] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): A_ = scores / self.temperature return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : str , UpperCAmelCase : float , UpperCAmelCase : float = -float("Inf" ) , UpperCAmelCase : int = 1 ): if not isinstance(UpperCAmelCase , UpperCAmelCase ) or (top_p < 0 or top_p > 1.0): raise ValueError(f'''`top_p` has to be a float > 0 and < 1, but is {top_p}''' ) if not isinstance(UpperCAmelCase , UpperCAmelCase ) or (min_tokens_to_keep < 1): raise ValueError(f'''`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}''' ) A_ = top_p A_ = filter_value A_ = min_tokens_to_keep def __call__( self : Union[str, Any] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): A_ , A_ = lax.top_k(UpperCAmelCase , scores.shape[-1] ) A_ = jnp.full_like(UpperCAmelCase , self.filter_value ) A_ = jax.nn.softmax(UpperCAmelCase , axis=-1 ).cumsum(axis=-1 ) A_ = cumulative_probs < self.top_p # include the token that is higher than top_p as well A_ = jnp.roll(UpperCAmelCase , 1 ) score_mask |= score_mask.at[:, 0].set(UpperCAmelCase ) # min tokens to keep A_ = score_mask.at[:, : self.min_tokens_to_keep].set(UpperCAmelCase ) A_ = jnp.where(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) A_ = jax.lax.sort_key_val(UpperCAmelCase , UpperCAmelCase )[-1] return next_scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : Any , UpperCAmelCase : int , UpperCAmelCase : float = -float("Inf" ) , UpperCAmelCase : int = 1 ): if not isinstance(UpperCAmelCase , UpperCAmelCase ) or top_k <= 0: raise ValueError(f'''`top_k` has to be a strictly positive integer, but is {top_k}''' ) A_ = max(UpperCAmelCase , UpperCAmelCase ) A_ = filter_value def __call__( self : List[str] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): A_ , A_ = scores.shape A_ = jnp.full(batch_size * vocab_size , self.filter_value ) A_ = min(self.top_k , scores.shape[-1] ) # Safety check A_ , A_ = lax.top_k(UpperCAmelCase , UpperCAmelCase ) A_ = jnp.broadcast_to((jnp.arange(UpperCAmelCase ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() A_ = topk_scores.flatten() A_ = topk_indices.flatten() + shift A_ = next_scores_flat.at[topk_indices_flat].set(UpperCAmelCase ) A_ = next_scores_flat.reshape(UpperCAmelCase , UpperCAmelCase ) return next_scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : Tuple , UpperCAmelCase : int ): A_ = bos_token_id def __call__( self : Tuple , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): A_ = jnp.full(scores.shape , -float("inf" ) ) A_ = 1 - jnp.bool_(cur_len - 1 ) A_ = jnp.where(UpperCAmelCase , new_scores.at[:, self.bos_token_id].set(0 ) , UpperCAmelCase ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : str , UpperCAmelCase : int , UpperCAmelCase : int ): A_ = max_length A_ = eos_token_id def __call__( self : Optional[int] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): A_ = jnp.full(scores.shape , -float("inf" ) ) A_ = 1 - jnp.bool_(cur_len - self.max_length + 1 ) A_ = jnp.where(UpperCAmelCase , new_scores.at[:, self.eos_token_id].set(0 ) , UpperCAmelCase ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : Optional[int] , UpperCAmelCase : int , UpperCAmelCase : int ): if not isinstance(UpperCAmelCase , UpperCAmelCase ) or min_length < 0: raise ValueError(f'''`min_length` has to be a positive integer, but is {min_length}''' ) if not isinstance(UpperCAmelCase , UpperCAmelCase ) or eos_token_id < 0: raise ValueError(f'''`eos_token_id` has to be a positive integer, but is {eos_token_id}''' ) A_ = min_length A_ = eos_token_id def __call__( self : Union[str, Any] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): # create boolean flag to decide if min length penalty should be applied A_ = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) A_ = jnp.where(UpperCAmelCase , scores.at[:, self.eos_token_id].set(-float("inf" ) ) , UpperCAmelCase ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : Optional[int] , UpperCAmelCase : int , UpperCAmelCase : Dict ): A_ = list(UpperCAmelCase ) A_ = begin_index def __call__( self : Union[str, Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : int , UpperCAmelCase : int ): A_ = 1 - jnp.bool_(cur_len - self.begin_index ) A_ = jnp.where(UpperCAmelCase , scores.at[:, self.begin_suppress_tokens].set(-float("inf" ) ) , UpperCAmelCase ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : Optional[int] , UpperCAmelCase : list ): A_ = list(UpperCAmelCase ) def __call__( self : Dict , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): A_ = scores.at[..., self.suppress_tokens].set(-float("inf" ) ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : List[str] , UpperCAmelCase : List[Any] ): A_ = dict(UpperCAmelCase ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. A_ = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: A_ = force_token_array.at[index].set(UpperCAmelCase ) A_ = jnp.intaa(UpperCAmelCase ) def __call__( self : Optional[int] , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : jnp.ndarray , UpperCAmelCase : int ): def _force_token(UpperCAmelCase : str ): A_ = scores.shape[0] A_ = self.force_token_array[generation_idx] A_ = jnp.ones_like(UpperCAmelCase , dtype=scores.dtype ) * -float("inf" ) A_ = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) A_ = lax.dynamic_update_slice(UpperCAmelCase , UpperCAmelCase , (0, current_token) ) return new_scores A_ = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(UpperCAmelCase ) , lambda: scores , ) , ) return scores class _a ( snake_case_ ): """simple docstring""" def __init__( self : str , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[Any] , UpperCAmelCase : Optional[int] ): A_ = generate_config.eos_token_id A_ = generate_config.no_timestamps_token_id A_ = generate_config.no_timestamps_token_id + 1 A_ = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(UpperCAmelCase , "max_initial_timestamp_index" ): A_ = generate_config.max_initial_timestamp_index else: A_ = model_config.vocab_size if self.max_initial_timestamp_index is None: A_ = model_config.vocab_size def __call__( self : Optional[Any] , UpperCAmelCase : Any , UpperCAmelCase : Tuple , UpperCAmelCase : str ): # suppress <|notimestamps|> which is handled by without_timestamps A_ = scores.at[:, self.no_timestamps_token_id].set(-float("inf" ) ) def handle_pairs(UpperCAmelCase : List[str] , UpperCAmelCase : Dict ): A_ = jnp.where((cur_len - self.begin_index) >= 1 , UpperCAmelCase , UpperCAmelCase ) A_ = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , UpperCAmelCase , ) A_ = jnp.where((cur_len - self.begin_index) < 2 , UpperCAmelCase , UpperCAmelCase ) A_ = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , UpperCAmelCase , UpperCAmelCase , ) return jnp.where( UpperCAmelCase , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float("inf" ) ) , scores_k.at[: self.eos_token_id].set(-float("inf" ) ) , ) , UpperCAmelCase , ) A_ = jax.vmap(UpperCAmelCase )(UpperCAmelCase , UpperCAmelCase ) A_ = jnp.where(cur_len == self.begin_index , UpperCAmelCase , UpperCAmelCase ) A_ = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , UpperCAmelCase , ) A_ = self.timestamp_begin + self.max_initial_timestamp_index A_ = jnp.where( UpperCAmelCase , scores.at[:, last_allowed + 1 :].set(-float("inf" ) ) , UpperCAmelCase , ) # if sum of probability over timestamps is above any other token, sample timestamp A_ = jax.nn.log_softmax(UpperCAmelCase , axis=-1 ) def handle_cumulative_probs(UpperCAmelCase : List[Any] , UpperCAmelCase : Optional[int] ): A_ = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) A_ = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float("inf" ) ) , UpperCAmelCase , ) A_ = jax.vmap(UpperCAmelCase )(UpperCAmelCase , UpperCAmelCase ) return scores
329
import math __a :Union[str, Any] = 10 __a :Union[str, Any] = 7 __a :int = BALLS_PER_COLOUR * NUM_COLOURS def __snake_case ( __UpperCamelCase : int = 20 ): """simple docstring""" A_ = math.comb(__UpperCamelCase ,__UpperCamelCase ) A_ = math.comb(NUM_BALLS - BALLS_PER_COLOUR ,__UpperCamelCase ) A_ = NUM_COLOURS * (1 - missing_colour / total) return f'''{result:.9f}''' if __name__ == "__main__": print(solution(20))
329
1
'''simple docstring''' import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoFeatureExtractor, WavaVecaFeatureExtractor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 lowercase_ = get_tests_dir("""fixtures""") class a_ ( unittest.TestCase ): '''simple docstring''' def snake_case_( self ) -> Union[str, Any]: # A mock response for an HTTP head request to emulate server down _SCREAMING_SNAKE_CASE = mock.Mock() _SCREAMING_SNAKE_CASE = 500 _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = HTTPError _SCREAMING_SNAKE_CASE = {} # Download this model to make sure it's in the cache. _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained("""hf-internal-testing/tiny-random-wav2vec2""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=A ) as mock_head: _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained("""hf-internal-testing/tiny-random-wav2vec2""" ) # This check we did call the fake head request mock_head.assert_called() def snake_case_( self ) -> str: # This test is for deprecated behavior and can be removed in v5 _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-wav2vec2/resolve/main/preprocessor_config.json""" ) @is_staging_test class a_ ( unittest.TestCase ): '''simple docstring''' @classmethod def snake_case_( cls ) -> Optional[int]: _SCREAMING_SNAKE_CASE = TOKEN HfFolder.save_token(A ) @classmethod def snake_case_( cls ) -> Dict: try: delete_repo(token=cls._token , repo_id="""test-feature-extractor""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-feature-extractor-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-feature-extractor""" ) except HTTPError: pass def snake_case_( self ) -> Optional[Any]: _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained(A ) feature_extractor.push_to_hub("""test-feature-extractor""" , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained(f'{USER}/test-feature-extractor' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-feature-extractor""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( A , repo_id="""test-feature-extractor""" , push_to_hub=A , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained(f'{USER}/test-feature-extractor' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained(A ) feature_extractor.push_to_hub("""valid_org/test-feature-extractor""" , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained("""valid_org/test-feature-extractor""" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-feature-extractor""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( A , repo_id="""valid_org/test-feature-extractor-org""" , push_to_hub=A , use_auth_token=self._token ) _SCREAMING_SNAKE_CASE = WavaVecaFeatureExtractor.from_pretrained("""valid_org/test-feature-extractor-org""" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(A , getattr(A , A ) ) def snake_case_( self ) -> Tuple: CustomFeatureExtractor.register_for_auto_class() _SCREAMING_SNAKE_CASE = CustomFeatureExtractor.from_pretrained(A ) feature_extractor.push_to_hub("""test-dynamic-feature-extractor""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( feature_extractor.auto_map , {"""AutoFeatureExtractor""": """custom_feature_extraction.CustomFeatureExtractor"""} , ) _SCREAMING_SNAKE_CASE = AutoFeatureExtractor.from_pretrained( f'{USER}/test-dynamic-feature-extractor' , trust_remote_code=A ) # Can't make an isinstance check because the new_feature_extractor is from the CustomFeatureExtractor class of a dynamic module self.assertEqual(new_feature_extractor.__class__.__name__ , """CustomFeatureExtractor""" )
58
'''simple docstring''' import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Accelerator def lowerCamelCase ( __lowerCamelCase : Tuple ) ->Tuple: _SCREAMING_SNAKE_CASE = fname.split(os.path.sep )[-1] return re.search(R"""^(.*)_\d+\.jpg$""" , __lowerCamelCase ).groups()[0] class a_ ( snake_case_ ): '''simple docstring''' def __init__( self , A , A=None , A=None ) -> int: _SCREAMING_SNAKE_CASE = file_names _SCREAMING_SNAKE_CASE = image_transform _SCREAMING_SNAKE_CASE = label_to_id def __len__( self ) -> Optional[Any]: return len(self.file_names ) def __getitem__( self , A ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = self.file_names[idx] _SCREAMING_SNAKE_CASE = PIL.Image.open(A ) _SCREAMING_SNAKE_CASE = raw_image.convert("""RGB""" ) if self.image_transform is not None: _SCREAMING_SNAKE_CASE = self.image_transform(A ) _SCREAMING_SNAKE_CASE = extract_label(A ) if self.label_to_id is not None: _SCREAMING_SNAKE_CASE = self.label_to_id[label] return {"image": image, "label": label} def lowerCamelCase ( __lowerCamelCase : Any , __lowerCamelCase : Tuple ) ->str: # Initialize accelerator if args.with_tracking: _SCREAMING_SNAKE_CASE = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="""all""" , project_dir=args.project_dir ) else: _SCREAMING_SNAKE_CASE = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _SCREAMING_SNAKE_CASE = config["""lr"""] _SCREAMING_SNAKE_CASE = int(config["""num_epochs"""] ) _SCREAMING_SNAKE_CASE = int(config["""seed"""] ) _SCREAMING_SNAKE_CASE = int(config["""batch_size"""] ) _SCREAMING_SNAKE_CASE = config["""image_size"""] if not isinstance(__lowerCamelCase , (list, tuple) ): _SCREAMING_SNAKE_CASE = (image_size, image_size) # Parse out whether we are saving every epoch or after a certain number of batches if hasattr(args.checkpointing_steps , """isdigit""" ): if args.checkpointing_steps == "epoch": _SCREAMING_SNAKE_CASE = args.checkpointing_steps elif args.checkpointing_steps.isdigit(): _SCREAMING_SNAKE_CASE = int(args.checkpointing_steps ) else: raise ValueError( F'Argument `checkpointing_steps` must be either a number or `epoch`. `{args.checkpointing_steps}` passed.' ) else: _SCREAMING_SNAKE_CASE = None # We need to initialize the trackers we use, and also store our configuration if args.with_tracking: _SCREAMING_SNAKE_CASE = os.path.split(__lowerCamelCase )[-1].split(""".""" )[0] accelerator.init_trackers(__lowerCamelCase , __lowerCamelCase ) # Grab all the image filenames _SCREAMING_SNAKE_CASE = [os.path.join(args.data_dir , __lowerCamelCase ) for fname in os.listdir(args.data_dir ) if fname.endswith(""".jpg""" )] # Build the label correspondences _SCREAMING_SNAKE_CASE = [extract_label(__lowerCamelCase ) for fname in file_names] _SCREAMING_SNAKE_CASE = list(set(__lowerCamelCase ) ) id_to_label.sort() _SCREAMING_SNAKE_CASE = {lbl: i for i, lbl in enumerate(__lowerCamelCase )} # Set the seed before splitting the data. np.random.seed(__lowerCamelCase ) torch.manual_seed(__lowerCamelCase ) torch.cuda.manual_seed_all(__lowerCamelCase ) # Split our filenames between train and validation _SCREAMING_SNAKE_CASE = np.random.permutation(len(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE = int(0.8 * len(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE = random_perm[:cut] _SCREAMING_SNAKE_CASE = random_perm[cut:] # For training we use a simple RandomResizedCrop _SCREAMING_SNAKE_CASE = Compose([RandomResizedCrop(__lowerCamelCase , scale=(0.5, 1.0) ), ToTensor()] ) _SCREAMING_SNAKE_CASE = PetsDataset( [file_names[i] for i in train_split] , image_transform=__lowerCamelCase , label_to_id=__lowerCamelCase ) # For evaluation, we use a deterministic Resize _SCREAMING_SNAKE_CASE = Compose([Resize(__lowerCamelCase ), ToTensor()] ) _SCREAMING_SNAKE_CASE = PetsDataset([file_names[i] for i in eval_split] , image_transform=__lowerCamelCase , label_to_id=__lowerCamelCase ) # Instantiate dataloaders. _SCREAMING_SNAKE_CASE = DataLoader(__lowerCamelCase , shuffle=__lowerCamelCase , batch_size=__lowerCamelCase , num_workers=4 ) _SCREAMING_SNAKE_CASE = DataLoader(__lowerCamelCase , shuffle=__lowerCamelCase , batch_size=__lowerCamelCase , num_workers=4 ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _SCREAMING_SNAKE_CASE = create_model("""resnet50d""" , pretrained=__lowerCamelCase , num_classes=len(__lowerCamelCase ) ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _SCREAMING_SNAKE_CASE = model.to(accelerator.device ) # Freezing the base model for param in model.parameters(): _SCREAMING_SNAKE_CASE = False for param in model.get_classifier().parameters(): _SCREAMING_SNAKE_CASE = True # We normalize the batches of images to be a bit faster. _SCREAMING_SNAKE_CASE = torch.tensor(model.default_cfg["""mean"""] )[None, :, None, None].to(accelerator.device ) _SCREAMING_SNAKE_CASE = torch.tensor(model.default_cfg["""std"""] )[None, :, None, None].to(accelerator.device ) # Instantiate optimizer _SCREAMING_SNAKE_CASE = torch.optim.Adam(params=model.parameters() , lr=lr / 25 ) # Instantiate learning rate scheduler _SCREAMING_SNAKE_CASE = OneCycleLR(optimizer=__lowerCamelCase , max_lr=__lowerCamelCase , epochs=__lowerCamelCase , steps_per_epoch=len(__lowerCamelCase ) ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = accelerator.prepare( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # We need to keep track of how many total steps we have iterated over _SCREAMING_SNAKE_CASE = 0 # We also need to keep track of the starting epoch so files are named properly _SCREAMING_SNAKE_CASE = 0 # Potentially load in the weights and states from a previous save if args.resume_from_checkpoint: if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "": accelerator.print(F'Resumed from checkpoint: {args.resume_from_checkpoint}' ) accelerator.load_state(args.resume_from_checkpoint ) _SCREAMING_SNAKE_CASE = os.path.basename(args.resume_from_checkpoint ) else: # Get the most recent checkpoint _SCREAMING_SNAKE_CASE = [f.name for f in os.scandir(os.getcwd() ) if f.is_dir()] dirs.sort(key=os.path.getctime ) _SCREAMING_SNAKE_CASE = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last # Extract `epoch_{i}` or `step_{i}` _SCREAMING_SNAKE_CASE = os.path.splitext(__lowerCamelCase )[0] if "epoch" in training_difference: _SCREAMING_SNAKE_CASE = int(training_difference.replace("""epoch_""" , """""" ) ) + 1 _SCREAMING_SNAKE_CASE = None else: _SCREAMING_SNAKE_CASE = int(training_difference.replace("""step_""" , """""" ) ) _SCREAMING_SNAKE_CASE = resume_step // len(__lowerCamelCase ) resume_step -= starting_epoch * len(__lowerCamelCase ) # Now we train the model for epoch in range(__lowerCamelCase , __lowerCamelCase ): model.train() if args.with_tracking: _SCREAMING_SNAKE_CASE = 0 if args.resume_from_checkpoint and epoch == starting_epoch and resume_step is not None: # We need to skip steps until we reach the resumed step _SCREAMING_SNAKE_CASE = accelerator.skip_first_batches(__lowerCamelCase , __lowerCamelCase ) overall_step += resume_step else: # After the first iteration though, we need to go back to the original dataloader _SCREAMING_SNAKE_CASE = train_dataloader for batch in active_dataloader: # We could avoid this line since we set the accelerator with `device_placement=True`. _SCREAMING_SNAKE_CASE = {k: v.to(accelerator.device ) for k, v in batch.items()} _SCREAMING_SNAKE_CASE = (batch["""image"""] - mean) / std _SCREAMING_SNAKE_CASE = model(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = torch.nn.functional.cross_entropy(__lowerCamelCase , batch["""label"""] ) # We keep track of the loss at each epoch if args.with_tracking: total_loss += loss.detach().float() accelerator.backward(__lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 if isinstance(__lowerCamelCase , __lowerCamelCase ): _SCREAMING_SNAKE_CASE = F'step_{overall_step}' if overall_step % checkpointing_steps == 0: if args.output_dir is not None: _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir , __lowerCamelCase ) accelerator.save_state(__lowerCamelCase ) model.eval() _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 0 for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. _SCREAMING_SNAKE_CASE = {k: v.to(accelerator.device ) for k, v in batch.items()} _SCREAMING_SNAKE_CASE = (batch["""image"""] - mean) / std with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(__lowerCamelCase ) _SCREAMING_SNAKE_CASE = outputs.argmax(dim=-1 ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = accelerator.gather_for_metrics((predictions, batch["""label"""]) ) _SCREAMING_SNAKE_CASE = predictions == references num_elems += accurate_preds.shape[0] accurate += accurate_preds.long().sum() _SCREAMING_SNAKE_CASE = accurate.item() / num_elems # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}: {100 * eval_metric:.2f}' ) if args.with_tracking: accelerator.log( { """accuracy""": 100 * eval_metric, """train_loss""": total_loss.item() / len(__lowerCamelCase ), """epoch""": epoch, } , step=__lowerCamelCase , ) if checkpointing_steps == "epoch": _SCREAMING_SNAKE_CASE = F'epoch_{epoch}' if args.output_dir is not None: _SCREAMING_SNAKE_CASE = os.path.join(args.output_dir , __lowerCamelCase ) accelerator.save_state(__lowerCamelCase ) if args.with_tracking: accelerator.end_training() def lowerCamelCase ( ) ->int: _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument("""--data_dir""" , required=__lowerCamelCase , help="""The data folder on disk.""" ) parser.add_argument("""--fp16""" , action="""store_true""" , help="""If passed, will use FP16 training.""" ) parser.add_argument( """--mixed_precision""" , type=__lowerCamelCase , default=__lowerCamelCase , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) parser.add_argument( """--checkpointing_steps""" , type=__lowerCamelCase , default=__lowerCamelCase , help="""Whether the various states should be saved at the end of every n steps, or 'epoch' for each epoch.""" , ) parser.add_argument( """--output_dir""" , type=__lowerCamelCase , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--resume_from_checkpoint""" , type=__lowerCamelCase , default=__lowerCamelCase , help="""If the training should continue from a checkpoint folder.""" , ) parser.add_argument( """--with_tracking""" , action="""store_true""" , help="""Whether to load in all available experiment trackers from the environment and use them for logging.""" , ) parser.add_argument( """--project_dir""" , type=__lowerCamelCase , default="""logs""" , help="""Location on where to store experiment tracking logs` and relevent project information""" , ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = {"""lr""": 3e-2, """num_epochs""": 3, """seed""": 42, """batch_size""": 64, """image_size""": 224} training_function(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": main()
58
1
'''simple docstring''' import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa UpperCamelCase__ = logging.getLogger(__name__) class lowerCamelCase_ ( __a ): lowerCAmelCase__ = 'summarization' lowerCAmelCase__ = ['loss'] lowerCAmelCase__ = ROUGE_KEYS lowerCAmelCase__ = 'rouge2' def __init__( self : str , _A : Optional[int] , **_A : Optional[Any] ): '''simple docstring''' if hparams.sortish_sampler and hparams.gpus > 1: UpperCAmelCase__ : Optional[int] = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError('''Dynamic Batch size does not work for multi-gpu training''' ) if hparams.sortish_sampler: raise ValueError('''--sortish_sampler and --max_tokens_per_batch may not be used simultaneously''' ) super().__init__(_A , num_labels=_A , mode=self.mode , **_A ) use_task_specific_params(self.model , '''summarization''' ) save_git_info(self.hparams.output_dir ) UpperCAmelCase__ : Optional[int] = Path(self.output_dir ) / '''metrics.json''' UpperCAmelCase__ : Tuple = Path(self.output_dir ) / '''hparams.pkl''' pickle_save(self.hparams , self.hparams_save_path ) UpperCAmelCase__ : Tuple = 0 UpperCAmelCase__ : Optional[int] = defaultdict(_A ) UpperCAmelCase__ : Optional[Any] = self.config.model_type UpperCAmelCase__ : int = self.config.tgt_vocab_size if self.model_type == '''fsmt''' else self.config.vocab_size UpperCAmelCase__ : dict = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } UpperCAmelCase__ : Dict = { '''train''': self.hparams.n_train, '''val''': self.hparams.n_val, '''test''': self.hparams.n_test, } UpperCAmelCase__ : str = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} UpperCAmelCase__ : Any = { '''train''': self.hparams.max_target_length, '''val''': self.hparams.val_max_target_length, '''test''': self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], f"""target_lens: {self.target_lens}""" assert self.target_lens["train"] <= self.target_lens["test"], f"""target_lens: {self.target_lens}""" if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) UpperCAmelCase__ : Tuple = get_git_info()['''repo_sha'''] UpperCAmelCase__ : List[str] = hparams.num_workers UpperCAmelCase__ : Tuple = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , _A ): UpperCAmelCase__ : Dict = self.tokenizer.lang_code_to_id[hparams.tgt_lang] UpperCAmelCase__ : Dict = self.decoder_start_token_id UpperCAmelCase__ : Union[str, Any] = ( SeqaSeqDataset if hasattr(self.tokenizer , '''prepare_seq2seq_batch''' ) else LegacySeqaSeqDataset ) UpperCAmelCase__ : Union[str, Any] = False UpperCAmelCase__ : Tuple = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: UpperCAmelCase__ : Optional[int] = self.hparams.eval_max_gen_length else: UpperCAmelCase__ : Union[str, Any] = self.model.config.max_length UpperCAmelCase__ : Tuple = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def lowercase_ ( self : List[str] , _A : Dict[str, torch.Tensor] ): '''simple docstring''' UpperCAmelCase__ : int = { k: self.tokenizer.batch_decode(v.tolist() ) if '''mask''' not in k else v.shape for k, v in batch.items() } save_json(_A , Path(self.output_dir ) / '''text_batch.json''' ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / '''tok_batch.json''' ) UpperCAmelCase__ : int = True return readable_batch def lowercase_ ( self : Union[str, Any] , _A : List[Any] , **_A : Union[str, Any] ): '''simple docstring''' return self.model(_A , **_A ) def lowercase_ ( self : str , _A : List[int] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.tokenizer.batch_decode( _A , skip_special_tokens=_A , clean_up_tokenization_spaces=_A ) return lmap(str.strip , _A ) def lowercase_ ( self : Optional[int] , _A : dict ): '''simple docstring''' UpperCAmelCase__ : Any = self.tokenizer.pad_token_id UpperCAmelCase__ : Any = batch['''input_ids'''], batch['''attention_mask'''] UpperCAmelCase__ : Any = batch['''labels'''] if isinstance(self.model , _A ): UpperCAmelCase__ : Tuple = self.model._shift_right(_A ) else: UpperCAmelCase__ : Dict = shift_tokens_right(_A , _A ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero UpperCAmelCase__ : int = decoder_input_ids self.save_readable_batch(_A ) UpperCAmelCase__ : List[str] = self(_A , attention_mask=_A , decoder_input_ids=_A , use_cache=_A ) UpperCAmelCase__ : Optional[Any] = outputs['''logits'''] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id UpperCAmelCase__ : Dict = nn.CrossEntropyLoss(ignore_index=_A ) assert lm_logits.shape[-1] == self.vocab_size UpperCAmelCase__ : int = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: UpperCAmelCase__ : Dict = nn.functional.log_softmax(_A , dim=-1 ) UpperCAmelCase__ : List[str] = label_smoothed_nll_loss( _A , _A , self.hparams.label_smoothing , ignore_index=_A ) return (loss,) @property def lowercase_ ( self : Tuple ): '''simple docstring''' return self.tokenizer.pad_token_id def lowercase_ ( self : Optional[int] , _A : List[Any] , _A : List[Any] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self._step(_A ) UpperCAmelCase__ : Optional[int] = dict(zip(self.loss_names , _A ) ) # tokens per batch UpperCAmelCase__ : Optional[int] = batch['''input_ids'''].ne(self.pad ).sum() + batch['''labels'''].ne(self.pad ).sum() UpperCAmelCase__ : Dict = batch['''input_ids'''].shape[0] UpperCAmelCase__ : Dict = batch['''input_ids'''].eq(self.pad ).sum() UpperCAmelCase__ : Union[str, Any] = batch['''input_ids'''].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def lowercase_ ( self : Tuple , _A : Dict , _A : Optional[int] ): '''simple docstring''' return self._generative_step(_A ) def lowercase_ ( self : str , _A : Optional[Any] , _A : Optional[Any]="val" ): '''simple docstring''' self.step_count += 1 UpperCAmelCase__ : Tuple = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} UpperCAmelCase__ : Optional[int] = losses['''loss'''] UpperCAmelCase__ : int = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ['''gen_time''', '''gen_len'''] } UpperCAmelCase__ : List[str] = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) UpperCAmelCase__ : torch.FloatTensor = torch.tensor(_A ).type_as(_A ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(_A ) UpperCAmelCase__ : str = {f"""{prefix}_avg_{k}""": x for k, x in losses.items()} UpperCAmelCase__ : Dict = self.step_count self.metrics[prefix].append(_A ) # callback writes this to self.metrics_save_path UpperCAmelCase__ : int = flatten_list([x['''preds'''] for x in outputs] ) return { "log": all_metrics, "preds": preds, f"""{prefix}_loss""": loss, f"""{prefix}_{self.val_metric}""": metric_tensor, } def lowercase_ ( self : Any , _A : List[Any] , _A : Tuple ): '''simple docstring''' return calculate_rouge(_A , _A ) def lowercase_ ( self : Dict , _A : dict ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') UpperCAmelCase__ : Tuple = self.model.generate( batch['''input_ids'''] , attention_mask=batch['''attention_mask'''] , use_cache=_A , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) UpperCAmelCase__ : List[Any] = (time.time() - ta) / batch['''input_ids'''].shape[0] UpperCAmelCase__ : List[str] = self.ids_to_clean_text(_A ) UpperCAmelCase__ : List[str] = self.ids_to_clean_text(batch['''labels'''] ) UpperCAmelCase__ : Any = self._step(_A ) UpperCAmelCase__ : List[str] = dict(zip(self.loss_names , _A ) ) UpperCAmelCase__ : Dict = self.calc_generative_metrics(_A , _A ) UpperCAmelCase__ : int = np.mean(lmap(_A , _A ) ) base_metrics.update(gen_time=_A , gen_len=_A , preds=_A , target=_A , **_A ) return base_metrics def lowercase_ ( self : List[str] , _A : str , _A : Dict ): '''simple docstring''' return self._generative_step(_A ) def lowercase_ ( self : List[str] , _A : Any ): '''simple docstring''' return self.validation_epoch_end(_A , prefix='''test''' ) def lowercase_ ( self : str , _A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : int = self.n_obs[type_path] UpperCAmelCase__ : Optional[int] = self.target_lens[type_path] UpperCAmelCase__ : List[Any] = self.dataset_class( self.tokenizer , type_path=_A , n_obs=_A , max_target_length=_A , **self.dataset_kwargs , ) return dataset def lowercase_ ( self : Optional[Any] , _A : str , _A : int , _A : bool = False ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.get_dataset(_A ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": UpperCAmelCase__ : Optional[Any] = dataset.make_sortish_sampler(_A , distributed=self.hparams.gpus > 1 ) return DataLoader( _A , batch_size=_A , collate_fn=dataset.collate_fn , shuffle=_A , num_workers=self.num_workers , sampler=_A , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": UpperCAmelCase__ : Union[str, Any] = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( _A , batch_sampler=_A , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( _A , batch_size=_A , collate_fn=dataset.collate_fn , shuffle=_A , num_workers=self.num_workers , sampler=_A , ) def lowercase_ ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self.get_dataloader('''train''' , batch_size=self.hparams.train_batch_size , shuffle=_A ) return dataloader def lowercase_ ( self : str ): '''simple docstring''' return self.get_dataloader('''val''' , batch_size=self.hparams.eval_batch_size ) def lowercase_ ( self : Dict ): '''simple docstring''' return self.get_dataloader('''test''' , batch_size=self.hparams.eval_batch_size ) @staticmethod def lowercase_ ( _A : Tuple , _A : List[Any] ): '''simple docstring''' BaseTransformer.add_model_specific_args(_A , _A ) add_generic_args(_A , _A ) parser.add_argument( '''--max_source_length''' , default=1_024 , type=_A , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--max_target_length''' , default=56 , type=_A , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--val_max_target_length''' , default=142 , type=_A , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--test_max_target_length''' , default=142 , type=_A , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument('''--freeze_encoder''' , action='''store_true''' ) parser.add_argument('''--freeze_embeds''' , action='''store_true''' ) parser.add_argument('''--sortish_sampler''' , action='''store_true''' , default=_A ) parser.add_argument('''--overwrite_output_dir''' , action='''store_true''' , default=_A ) parser.add_argument('''--max_tokens_per_batch''' , type=_A , default=_A ) parser.add_argument('''--logger_name''' , type=_A , choices=['''default''', '''wandb''', '''wandb_shared'''] , default='''default''' ) parser.add_argument('''--n_train''' , type=_A , default=-1 , required=_A , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_val''' , type=_A , default=500 , required=_A , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_test''' , type=_A , default=-1 , required=_A , help='''# examples. -1 means use all.''' ) parser.add_argument( '''--task''' , type=_A , default='''summarization''' , required=_A , help='''# examples. -1 means use all.''' ) parser.add_argument('''--label_smoothing''' , type=_A , default=0.0 , required=_A ) parser.add_argument('''--src_lang''' , type=_A , default='''''' , required=_A ) parser.add_argument('''--tgt_lang''' , type=_A , default='''''' , required=_A ) parser.add_argument('''--eval_beams''' , type=_A , default=_A , required=_A ) parser.add_argument( '''--val_metric''' , type=_A , default=_A , required=_A , choices=['''bleu''', '''rouge2''', '''loss''', None] ) parser.add_argument('''--eval_max_gen_length''' , type=_A , default=_A , help='''never generate more than n tokens''' ) parser.add_argument('''--save_top_k''' , type=_A , default=1 , required=_A , help='''How many checkpoints to save''' ) parser.add_argument( '''--early_stopping_patience''' , type=_A , default=-1 , required=_A , help=( '''-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So''' ''' val_check_interval will effect it.''' ) , ) return parser class lowerCamelCase_ ( __a ): lowerCAmelCase__ = 'translation' lowerCAmelCase__ = ['loss'] lowerCAmelCase__ = ['bleu'] lowerCAmelCase__ = 'bleu' def __init__( self : List[str] , _A : Tuple , **_A : Optional[Any] ): '''simple docstring''' super().__init__(_A , **_A ) UpperCAmelCase__ : Optional[Any] = hparams.src_lang UpperCAmelCase__ : Any = hparams.tgt_lang def lowercase_ ( self : Dict , _A : Any , _A : Any ): '''simple docstring''' return calculate_bleu(_A , _A ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__=None ): Path(args.output_dir ).mkdir(exist_ok=lowerCAmelCase__ ) check_output_dir(lowerCAmelCase__ , expected_items=3 ) if model is None: if "summarization" in args.task: UpperCAmelCase__ : SummarizationModule = SummarizationModule(lowerCAmelCase__ ) else: UpperCAmelCase__ : SummarizationModule = TranslationModule(lowerCAmelCase__ ) UpperCAmelCase__ : str = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith('''/tmp''' ) or str(args.output_dir ).startswith('''/var''' ) ): UpperCAmelCase__ : Optional[int] = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger UpperCAmelCase__ : Tuple = os.environ.get('''WANDB_PROJECT''' , lowerCAmelCase__ ) UpperCAmelCase__ : List[str] = WandbLogger(name=model.output_dir.name , project=lowerCAmelCase__ ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger UpperCAmelCase__ : Dict = WandbLogger(name=model.output_dir.name , project=F"""hf_{dataset}""" ) if args.early_stopping_patience >= 0: UpperCAmelCase__ : List[Any] = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : int = args.val_metric == '''loss''' UpperCAmelCase__ : pl.Trainer = generic_train( lowerCAmelCase__ , lowerCAmelCase__ , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , lowerCAmelCase__ ) , early_stopping_callback=lowerCAmelCase__ , logger=lowerCAmelCase__ , ) pickle_save(model.hparams , model.output_dir / '''hparams.pkl''' ) if not args.do_predict: return model UpperCAmelCase__ : str = '''''' UpperCAmelCase__ : Optional[Any] = sorted(glob.glob(os.path.join(args.output_dir , '''*.ckpt''' ) , recursive=lowerCAmelCase__ ) ) if checkpoints: UpperCAmelCase__ : List[Any] = checkpoints[-1] UpperCAmelCase__ : List[str] = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() UpperCamelCase__ = pl.Trainer.add_argparse_args(parser) UpperCamelCase__ = SummarizationModule.add_model_specific_args(parser, os.getcwd()) UpperCamelCase__ = parser.parse_args() main(args)
367
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def a__ ( lowerCAmelCase__ ) -> List[Any]: return 1 / (1 + np.exp(-z )) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: return (-y * np.log(lowerCAmelCase__ ) - (1 - y) * np.log(1 - h )).mean() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: UpperCAmelCase__ : str = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) return np.sum(y * scores - np.log(1 + np.exp(lowerCAmelCase__ ) ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=7_00_00 ) -> List[Any]: UpperCAmelCase__ : Tuple = np.zeros(x.shape[1] ) for iterations in range(lowerCAmelCase__ ): UpperCAmelCase__ : List[Any] = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase__ : List[str] = sigmoid_function(lowerCAmelCase__ ) UpperCAmelCase__ : int = np.dot(x.T , h - y ) / y.size UpperCAmelCase__ : Optional[int] = theta - alpha * gradient # updating the weights UpperCAmelCase__ : Dict = np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase__ : int = sigmoid_function(lowerCAmelCase__ ) UpperCAmelCase__ : Tuple = cost_function(lowerCAmelCase__ , lowerCAmelCase__ ) if iterations % 1_00 == 0: print(F"""loss: {j} \t""" ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": UpperCamelCase__ = datasets.load_iris() UpperCamelCase__ = iris.data[:, :2] UpperCamelCase__ = (iris.target != 0) * 1 UpperCamelCase__ = 0.1 UpperCamelCase__ = logistic_reg(alpha, x, y, max_iterations=7_0_0_0_0) print('''theta: ''', theta) # printing the theta i.e our weights vector def a__ ( lowerCAmelCase__ ) -> Dict: return sigmoid_function( np.dot(lowerCAmelCase__ , lowerCAmelCase__ ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(1_0, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((UpperCamelCase__) , (UpperCamelCase__)) = (x[:, 0].min(), x[:, 0].max()) ((UpperCamelCase__) , (UpperCamelCase__)) = (x[:, 1].min(), x[:, 1].max()) ((UpperCamelCase__) , (UpperCamelCase__)) = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) UpperCamelCase__ = np.c_[xxa.ravel(), xxa.ravel()] UpperCamelCase__ = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
299
0
"""simple docstring""" import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class a ( __snake_case, unittest.TestCase ): # TODO: is there an appropriate internal test set? UpperCAmelCase_ : int ="""ssube/stable-diffusion-x4-upscaler-onnx""" def UpperCamelCase_ ( self , _lowerCamelCase=0 ): lowercase = floats_tensor((1, 3, 1_2_8, 1_2_8) , rng=random.Random(_lowerCamelCase ) ) lowercase = torch.manual_seed(_lowerCamelCase ) lowercase = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase_ ( self ): lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = self.get_dummy_inputs() lowercase = pipe(**_lowerCamelCase ).images lowercase = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def UpperCamelCase_ ( self ): lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) lowercase = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = self.get_dummy_inputs() lowercase = pipe(**_lowerCamelCase ).images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array( [0.6_8_9_8_8_9_2, 0.5_9_2_4_0_5_5_6, 0.5_2_4_9_9_5_2_7, 0.5_8_8_6_6_2_1_5, 0.5_2_2_5_8_2_3_5, 0.5_2_5_7_2_7_1_5, 0.6_2_4_1_4_4_7_3, 0.6_1_7_4_3_8_7, 0.6_2_1_4_9_6_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCamelCase_ ( self ): lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) lowercase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = self.get_dummy_inputs() lowercase = pipe(**_lowerCamelCase ).images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array( [0.7_6_5_9_2_7_8, 0.7_6_4_3_7_6_6_4, 0.7_5_5_7_9_1_0_7, 0.7_6_9_1_1_1_6, 0.7_7_6_6_6_9_8_6, 0.7_7_2_7_6_7_2, 0.7_7_5_8_6_6_4, 0.7_8_1_2_2_2_6, 0.7_6_9_4_2_5_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCamelCase_ ( self ): lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) lowercase = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = self.get_dummy_inputs() lowercase = pipe(**_lowerCamelCase ).images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCamelCase_ ( self ): lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='CPUExecutionProvider' ) lowercase = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = self.get_dummy_inputs() lowercase = pipe(**_lowerCamelCase ).images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array( [0.7_7_4_2_4_4_9_6, 0.7_7_3_6_0_1, 0.7_6_4_5_2_8_8, 0.7_7_6_9_5_9_8, 0.7_7_7_2_7_3_9, 0.7_7_3_8_6_8_8, 0.7_8_1_8_7_2_3_3, 0.7_7_8_7_9_5_8_4, 0.7_6_7_0_4_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class a ( unittest.TestCase ): @property def UpperCamelCase_ ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase_ ( self ): lowercase = ort.SessionOptions() lowercase = False return options def UpperCamelCase_ ( self ): lowercase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg' ) lowercase = init_image.resize((1_2_8, 1_2_8) ) # using the PNDM scheduler by default lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = 'A fantasy landscape, trending on artstation' lowercase = torch.manual_seed(0 ) lowercase = pipe( prompt=_lowerCamelCase , image=_lowerCamelCase , guidance_scale=7.5 , num_inference_steps=1_0 , generator=_lowerCamelCase , output_type='np' , ) lowercase = output.images lowercase = images[0, 2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert images.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array([0.4_8_8_3, 0.4_9_4_7, 0.4_9_8_0, 0.4_9_7_5, 0.4_9_8_2, 0.4_9_8_0, 0.5_0_0_0, 0.5_0_0_6, 0.4_9_7_2] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCamelCase_ ( self ): lowercase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg' ) lowercase = init_image.resize((1_2_8, 1_2_8) ) lowercase = LMSDiscreteScheduler.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' , subfolder='scheduler' ) lowercase = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx' , scheduler=_lowerCamelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) lowercase = 'A fantasy landscape, trending on artstation' lowercase = torch.manual_seed(0 ) lowercase = pipe( prompt=_lowerCamelCase , image=_lowerCamelCase , guidance_scale=7.5 , num_inference_steps=2_0 , generator=_lowerCamelCase , output_type='np' , ) lowercase = output.images lowercase = images[0, 2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert images.shape == (1, 5_1_2, 5_1_2, 3) lowercase = np.array( [0.5_0_1_7_3_7_5_3, 0.5_0_2_2_3_3_5_6, 0.5_0_2_0_3_9, 0.5_0_2_3_3_0_3_6, 0.5_0_2_3_7_2_5, 0.5_0_2_2_6_0_1, 0.5_0_1_8_7_5_8, 0.5_0_2_3_4_0_8_5, 0.5_0_2_4_1_5_6_6] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
220
import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal snake_case_ = datasets.utils.logging.get_logger(__name__) snake_case_ = ['''names''', '''prefix'''] snake_case_ = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] snake_case_ = ['''encoding_errors''', '''on_bad_lines'''] snake_case_ = ['''date_format'''] @dataclass class SCREAMING_SNAKE_CASE__ (datasets.BuilderConfig ): __lowerCamelCase : str = "," __lowerCamelCase : Optional[str] = None __lowerCamelCase : Optional[Union[int, List[int], str]] = "infer" __lowerCamelCase : Optional[List[str]] = None __lowerCamelCase : Optional[List[str]] = None __lowerCamelCase : Optional[Union[int, str, List[int], List[str]]] = None __lowerCamelCase : Optional[Union[List[int], List[str]]] = None __lowerCamelCase : Optional[str] = None __lowerCamelCase : bool = True __lowerCamelCase : Optional[Literal["c", "python", "pyarrow"]] = None __lowerCamelCase : Dict[Union[int, str], Callable[[Any], Any]] = None __lowerCamelCase : Optional[list] = None __lowerCamelCase : Optional[list] = None __lowerCamelCase : bool = False __lowerCamelCase : Optional[Union[int, List[int]]] = None __lowerCamelCase : Optional[int] = None __lowerCamelCase : Optional[Union[str, List[str]]] = None __lowerCamelCase : bool = True __lowerCamelCase : bool = True __lowerCamelCase : bool = False __lowerCamelCase : bool = True __lowerCamelCase : Optional[str] = None __lowerCamelCase : str = "." __lowerCamelCase : Optional[str] = None __lowerCamelCase : str = '"' __lowerCamelCase : int = 0 __lowerCamelCase : Optional[str] = None __lowerCamelCase : Optional[str] = None __lowerCamelCase : Optional[str] = None __lowerCamelCase : Optional[str] = None __lowerCamelCase : bool = True __lowerCamelCase : bool = True __lowerCamelCase : int = 0 __lowerCamelCase : bool = True __lowerCamelCase : bool = False __lowerCamelCase : Optional[str] = None __lowerCamelCase : int = 1_0000 __lowerCamelCase : Optional[datasets.Features] = None __lowerCamelCase : Optional[str] = "strict" __lowerCamelCase : Literal["error", "warn", "skip"] = "error" __lowerCamelCase : Optional[str] = None def snake_case_ ( self): if self.delimiter is not None: lowercase__ : List[Any] = self.delimiter if self.column_names is not None: lowercase__ : Optional[int] = self.column_names @property def snake_case_ ( self): lowercase__ : Dict = { 'sep': self.sep, 'header': self.header, 'names': self.names, 'index_col': self.index_col, 'usecols': self.usecols, 'prefix': self.prefix, 'mangle_dupe_cols': self.mangle_dupe_cols, 'engine': self.engine, 'converters': self.converters, 'true_values': self.true_values, 'false_values': self.false_values, 'skipinitialspace': self.skipinitialspace, 'skiprows': self.skiprows, 'nrows': self.nrows, 'na_values': self.na_values, 'keep_default_na': self.keep_default_na, 'na_filter': self.na_filter, 'verbose': self.verbose, 'skip_blank_lines': self.skip_blank_lines, 'thousands': self.thousands, 'decimal': self.decimal, 'lineterminator': self.lineterminator, 'quotechar': self.quotechar, 'quoting': self.quoting, 'escapechar': self.escapechar, 'comment': self.comment, 'encoding': self.encoding, 'dialect': self.dialect, 'error_bad_lines': self.error_bad_lines, 'warn_bad_lines': self.warn_bad_lines, 'skipfooter': self.skipfooter, 'doublequote': self.doublequote, 'memory_map': self.memory_map, 'float_precision': self.float_precision, 'chunksize': self.chunksize, 'encoding_errors': self.encoding_errors, 'on_bad_lines': self.on_bad_lines, 'date_format': self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , a): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class SCREAMING_SNAKE_CASE__ (datasets.ArrowBasedBuilder ): __lowerCamelCase : Optional[Any] = CsvConfig def snake_case_ ( self): return datasets.DatasetInfo(features=self.config.features) def snake_case_ ( self , a): if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""") lowercase__ : Any = dl_manager.download_and_extract(self.config.data_files) if isinstance(a , (str, list, tuple)): lowercase__ : List[str] = data_files if isinstance(a , a): lowercase__ : Optional[Any] = [files] lowercase__ : Optional[int] = [dl_manager.iter_files(a) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files})] lowercase__ : int = [] for split_name, files in data_files.items(): if isinstance(a , a): lowercase__ : Optional[int] = [files] lowercase__ : Tuple = [dl_manager.iter_files(a) for file in files] splits.append(datasets.SplitGenerator(name=a , gen_kwargs={'files': files})) return splits def snake_case_ ( self , a): if self.config.features is not None: lowercase__ : Optional[int] = self.config.features.arrow_schema if all(not require_storage_cast(a) for feature in self.config.features.values()): # cheaper cast lowercase__ : Dict = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=a) else: # more expensive cast; allows str <-> int/float or str to Audio for example lowercase__ : Optional[Any] = table_cast(a , a) return pa_table def snake_case_ ( self , a): lowercase__ : List[Any] = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str lowercase__ : Optional[int] = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(a) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values()) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(a)): lowercase__ : int = pd.read_csv(a , iterator=a , dtype=a , **self.config.pd_read_csv_kwargs) try: for batch_idx, df in enumerate(a): lowercase__ : List[str] = pa.Table.from_pandas(a) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(a) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(a)}: {e}""") raise
214
0
def lowerCamelCase__ ( snake_case_ : int = 1000 ) -> int: return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
356
# Algorithm for the pigeonhole sorting def lowerCamelCase__ ( snake_case_ : int ) -> Optional[int]: __snake_case = min(snake_case_ ) # min() finds the minimum value __snake_case = max(snake_case_ ) # max() finds the maximum value __snake_case = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size __snake_case = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(snake_case_ , snake_case_ ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. __snake_case = 0 for count in range(snake_case_ ): while holes[count] > 0: holes[count] -= 1 __snake_case = count + min_val i += 1 def lowerCamelCase__ ( ) -> Union[str, Any]: __snake_case = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(snake_case_ ) print('''Sorted order is:''' , ''' '''.join(snake_case_ ) ) if __name__ == "__main__": main()
238
0
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer __lowerCAmelCase : Dict = ['gpt2'] __lowerCAmelCase : Optional[Any] = 'gpt2' if is_tf_available(): class UpperCAmelCase_ ( tf.Module ): '''simple docstring''' def __init__( self : Union[str, Any] , UpperCamelCase__ : List[str] ) -> Any: """simple docstring""" super().__init__() __magic_name__ = tokenizer __magic_name__ = AutoConfig.from_pretrained(UpperCamelCase__ ) __magic_name__ = TFGPTaLMHeadModel.from_config(UpperCamelCase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) , tf.string , name="""text""" ),) ) def _lowercase ( self : List[str] , UpperCamelCase__ : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __magic_name__ = self.tokenizer(UpperCamelCase__ ) __magic_name__ = tokenized["""input_ids"""].to_tensor() __magic_name__ = tf.cast(input_ids_dense > 0 , tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __magic_name__ = self.model(input_ids=UpperCamelCase__ , attention_mask=UpperCamelCase__ )["""logits"""] return outputs @require_tf @require_keras_nlp class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowercase ( self : Dict ) -> Optional[Any]: """simple docstring""" super().setUp() __magic_name__ = [GPTaTokenizer.from_pretrained(UpperCamelCase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __magic_name__ = [TFGPTaTokenizer.from_pretrained(UpperCamelCase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __magic_name__ = [ """This is a straightforward English test sentence.""", """This one has some weird characters\rto\nsee\r\nif those\u00E9break things.""", """Now we're going to add some Chinese: 一 二 三 一二三""", """And some much more rare Chinese: 齉 堃 齉堃""", """Je vais aussi écrire en français pour tester les accents""", """Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ""", ] __magic_name__ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def _lowercase ( self : Any ) -> str: """simple docstring""" for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in self.test_sentences: __magic_name__ = tokenizer([test_inputs] , return_tensors="""tf""" ) __magic_name__ = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __magic_name__ = python_outputs[key].numpy() __magic_name__ = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(UpperCamelCase__ , tf.intaa ) == tf_outputs_values ) ) @slow def _lowercase ( self : Any ) -> Union[str, Any]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: __magic_name__ = tf.function(UpperCamelCase__ ) for test_inputs in self.test_sentences: __magic_name__ = tf.constant(UpperCamelCase__ ) __magic_name__ = compiled_tokenizer(UpperCamelCase__ ) __magic_name__ = tf_tokenizer(UpperCamelCase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: __magic_name__ = ModelToSave(tokenizer=UpperCamelCase__ ) __magic_name__ = tf.convert_to_tensor([self.test_sentences[0]] ) __magic_name__ = model.serving(UpperCamelCase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __magic_name__ = Path(UpperCamelCase__ ) / """saved.model""" tf.saved_model.save(UpperCamelCase__ , UpperCamelCase__ , signatures={"""serving_default""": model.serving} ) __magic_name__ = tf.saved_model.load(UpperCamelCase__ ) __magic_name__ = loaded_model.signatures["""serving_default"""](UpperCamelCase__ )["""output_0"""] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def _lowercase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: __magic_name__ = tf.convert_to_tensor([self.test_sentences[0]] ) __magic_name__ = tf_tokenizer(UpperCamelCase__ ) # Build model with some sample inputs __magic_name__ = tf_tokenizer.get_config() __magic_name__ = TFGPTaTokenizer.from_config(UpperCamelCase__ ) __magic_name__ = model_from_config(UpperCamelCase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def _lowercase ( self : Dict ) -> Any: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: # for the test to run __magic_name__ = 12_3123 for max_length in [3, 5, 1024]: __magic_name__ = tf.convert_to_tensor([self.test_sentences[0]] ) __magic_name__ = tf_tokenizer(UpperCamelCase__ , max_length=UpperCamelCase__ ) __magic_name__ = out["""input_ids"""].numpy().shape[1] assert out_length == max_length
88
from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class __UpperCAmelCase : __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : torch.Tensor # [batch_size x 3] __snake_case : int __snake_case : int __snake_case : float __snake_case : float __snake_case : Tuple[int] def UpperCamelCase ( self: str ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def UpperCamelCase ( self: List[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = torch.arange(self.height * self.width ) _SCREAMING_SNAKE_CASE = torch.stack( [ pixel_indices % self.width, torch.div(UpperCAmelCase_ , self.width , rounding_mode="""trunc""" ), ] , axis=1 , ) return coords @property def UpperCamelCase ( self: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE = self.shape _SCREAMING_SNAKE_CASE = int(np.prod(UpperCAmelCase_ ) ) _SCREAMING_SNAKE_CASE = self.get_image_coords() _SCREAMING_SNAKE_CASE = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) _SCREAMING_SNAKE_CASE = self.get_camera_rays(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = rays.view(UpperCAmelCase_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def UpperCamelCase ( self: Any , UpperCAmelCase_: torch.Tensor ): '''simple docstring''' _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] _SCREAMING_SNAKE_CASE = coords.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = self.resolution() _SCREAMING_SNAKE_CASE = self.fov() _SCREAMING_SNAKE_CASE = (flat.float() / (res - 1)) * 2 - 1 _SCREAMING_SNAKE_CASE = fracs * torch.tan(fov / 2 ) _SCREAMING_SNAKE_CASE = fracs.view(UpperCAmelCase_ , -1 , 2 ) _SCREAMING_SNAKE_CASE = ( self.z.view(UpperCAmelCase_ , 1 , 3 ) + self.x.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(UpperCAmelCase_ , 1 , 3 ) * fracs[:, :, 1:] ) _SCREAMING_SNAKE_CASE = directions / directions.norm(dim=-1 , keepdim=UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.stack( [ torch.broadcast_to(self.origin.view(UpperCAmelCase_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(UpperCAmelCase_ , *UpperCAmelCase_ , 2 , 3 ) def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: int , UpperCAmelCase_: int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=UpperCAmelCase_ , height=UpperCAmelCase_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __lowerCamelCase ( snake_case__ ) -> DifferentiableProjectiveCamera: """simple docstring""" _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = [] for theta in np.linspace(0 ,2 * np.pi ,num=20 ): _SCREAMING_SNAKE_CASE = np.array([np.sin(snake_case__ ), np.cos(snake_case__ ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) _SCREAMING_SNAKE_CASE = -z * 4 _SCREAMING_SNAKE_CASE = np.array([np.cos(snake_case__ ), -np.sin(snake_case__ ), 0.0] ) _SCREAMING_SNAKE_CASE = np.cross(snake_case__ ,snake_case__ ) origins.append(snake_case__ ) xs.append(snake_case__ ) ys.append(snake_case__ ) zs.append(snake_case__ ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,x=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,y=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,z=torch.from_numpy(np.stack(snake_case__ ,axis=0 ) ).float() ,width=snake_case__ ,height=snake_case__ ,x_fov=0.7 ,y_fov=0.7 ,shape=(1, len(snake_case__ )) ,)
306
0
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING a__ : Tuple = logging.get_logger(__name__) a__ : Any = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Optional[int] = "blip_2_vision_model" def __init__( self : Union[str, Any] , UpperCAmelCase__ : Optional[int]=1_4_0_8 , UpperCAmelCase__ : str=6_1_4_4 , UpperCAmelCase__ : int=3_9 , UpperCAmelCase__ : List[Any]=1_6 , UpperCAmelCase__ : Tuple=2_2_4 , UpperCAmelCase__ : Any=1_4 , UpperCAmelCase__ : List[str]="gelu" , UpperCAmelCase__ : Union[str, Any]=0.00_001 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : Dict=1E-10 , UpperCAmelCase__ : Dict=True , **UpperCAmelCase__ : List[Any] , ) -> List[str]: super().__init__(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = patch_size __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = attention_dropout __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = qkv_bias @classmethod def UpperCAmelCase_ ( cls : List[Any] , UpperCAmelCase__ : Union[str, os.PathLike] , **UpperCAmelCase__ : Union[str, Any] ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = cls.get_config_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": __SCREAMING_SNAKE_CASE = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : int = "blip_2_qformer" def __init__( self : Dict , UpperCAmelCase__ : List[str]=3_0_5_2_2 , UpperCAmelCase__ : Union[str, Any]=7_6_8 , UpperCAmelCase__ : Optional[Any]=1_2 , UpperCAmelCase__ : Dict=1_2 , UpperCAmelCase__ : List[Any]=3_0_7_2 , UpperCAmelCase__ : List[str]="gelu" , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : List[Any]=5_1_2 , UpperCAmelCase__ : str=0.02 , UpperCAmelCase__ : Any=1E-12 , UpperCAmelCase__ : Tuple=0 , UpperCAmelCase__ : int="absolute" , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : Tuple=1_4_0_8 , **UpperCAmelCase__ : Tuple , ) -> List[Any]: super().__init__(pad_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = position_embedding_type __SCREAMING_SNAKE_CASE = cross_attention_frequency __SCREAMING_SNAKE_CASE = encoder_hidden_size @classmethod def UpperCAmelCase_ ( cls : Union[str, Any] , UpperCAmelCase__ : Union[str, os.PathLike] , **UpperCAmelCase__ : List[str] ) -> "PretrainedConfig": cls._set_token_in_kwargs(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = cls.get_config_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": __SCREAMING_SNAKE_CASE = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(UpperCAmelCase__ , **UpperCAmelCase__ ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : Dict = "blip-2" snake_case__ : Dict = True def __init__( self : str , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Any=3_2 , **UpperCAmelCase__ : Union[str, Any] ) -> Tuple: super().__init__(**UpperCAmelCase__ ) if vision_config is None: __SCREAMING_SNAKE_CASE = {} logger.info("vision_config is None. initializing the Blip2VisionConfig with default values." ) if qformer_config is None: __SCREAMING_SNAKE_CASE = {} logger.info("qformer_config is None. Initializing the Blip2QFormerConfig with default values." ) if text_config is None: __SCREAMING_SNAKE_CASE = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) __SCREAMING_SNAKE_CASE = BlipaVisionConfig(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = BlipaQFormerConfig(**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = text_config["model_type"] if "model_type" in text_config else "opt" __SCREAMING_SNAKE_CASE = CONFIG_MAPPING[text_model_type](**UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = self.text_config.tie_word_embeddings __SCREAMING_SNAKE_CASE = self.text_config.is_encoder_decoder __SCREAMING_SNAKE_CASE = num_query_tokens __SCREAMING_SNAKE_CASE = self.vision_config.hidden_size __SCREAMING_SNAKE_CASE = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __SCREAMING_SNAKE_CASE = 1.0 __SCREAMING_SNAKE_CASE = 0.02 @classmethod def UpperCAmelCase_ ( cls : Optional[int] , UpperCAmelCase__ : BlipaVisionConfig , UpperCAmelCase__ : BlipaQFormerConfig , UpperCAmelCase__ : PretrainedConfig , **UpperCAmelCase__ : Union[str, Any] , ) -> Union[str, Any]: return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **UpperCAmelCase__ , ) def UpperCAmelCase_ ( self : Any ) -> Optional[int]: __SCREAMING_SNAKE_CASE = copy.deepcopy(self.__dict__ ) __SCREAMING_SNAKE_CASE = self.vision_config.to_dict() __SCREAMING_SNAKE_CASE = self.qformer_config.to_dict() __SCREAMING_SNAKE_CASE = self.text_config.to_dict() __SCREAMING_SNAKE_CASE = self.__class__.model_type return output
363
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
195
0
"""simple docstring""" def UpperCAmelCase__ ( lowerCAmelCase__ :float , lowerCAmelCase__ :float , lowerCAmelCase__ :int ) -> float: '''simple docstring''' if principal <= 0: raise Exception("""Principal borrowed must be > 0""" ) if rate_per_annum < 0: raise Exception("""Rate of interest must be >= 0""" ) if years_to_repay <= 0 or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise Exception("""Years to repay must be an integer > 0""" ) # Yearly rate is divided by 12 to get monthly rate lowercase = rate_per_annum / 1_2 # Years to repay is multiplied by 12 to get number of payments as payment is monthly lowercase = years_to_repay * 1_2 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
197
"""simple docstring""" from ..utils import DummyObject, requires_backends class _A ( metaclass=lowerCAmelCase ): snake_case__ : str = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : List[Any] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : Optional[Any] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : int = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : List[str] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=lowerCAmelCase ): snake_case__ : Optional[Any] = ['torch', 'transformers', 'onnx'] def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def A__ ( cls , *__lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] )
197
1
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import evaluate import numpy as np from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/text-classification/requirements.txt") lowerCamelCase : int = logging.getLogger(__name__) @dataclass class A__ : A__ = field( default=1_28 , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) A__ = field( default=A__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) A__ = field( default=A__ , metadata={ 'help': ( 'Whether to pad all samples to `max_seq_length`. ' 'If False, will pad the samples dynamically when batching to the maximum length in the batch.' ) } , ) A__ = field( default=A__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) A__ = field( default=A__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) A__ = field( default=A__ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of prediction examples to this ' 'value if set.' ) } , ) @dataclass class A__ : A__ = field( default=A__ , metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) A__ = field( default=A__ , metadata={'help': 'Evaluation language. Also train language if `train_language` is set to None.'} ) A__ = field( default=A__ , metadata={'help': 'Train language if it is different from the evaluation language.'} ) A__ = field( default=A__ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) A__ = field( default=A__ , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) A__ = field( default=A__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) A__ = field( default=A__ , metadata={'help': 'arg to indicate if tokenizer should do lower case in AutoTokenizer.from_pretrained()'} , ) A__ = field( default=A__ , metadata={'help': 'Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.'} , ) A__ = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) A__ = field( default=A__ , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) A__ = field( default=A__ , metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'} , ) def _lowerCAmelCase ( ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_xnli' , _UpperCamelCase ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _SCREAMING_SNAKE_CASE =training_args.get_process_log_level() logger.setLevel(_UpperCamelCase ) datasets.utils.logging.set_verbosity(_UpperCamelCase ) transformers.utils.logging.set_verbosity(_UpperCamelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}" + f"distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}" ) logger.info(f"Training/evaluation parameters {training_args}" ) # Detecting last checkpoint. _SCREAMING_SNAKE_CASE =None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _SCREAMING_SNAKE_CASE =get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"Output directory ({training_args.output_dir}) already exists and is not empty. " 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None: logger.info( f"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change " 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # In distributed training, the load_dataset function guarantees that only one local process can concurrently # download the dataset. # Downloading and loading xnli dataset from the hub. if training_args.do_train: if model_args.train_language is None: _SCREAMING_SNAKE_CASE =load_dataset( 'xnli' , model_args.language , split='train' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: _SCREAMING_SNAKE_CASE =load_dataset( 'xnli' , model_args.train_language , split='train' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) _SCREAMING_SNAKE_CASE =train_dataset.features['label'].names if training_args.do_eval: _SCREAMING_SNAKE_CASE =load_dataset( 'xnli' , model_args.language , split='validation' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) _SCREAMING_SNAKE_CASE =eval_dataset.features['label'].names if training_args.do_predict: _SCREAMING_SNAKE_CASE =load_dataset( 'xnli' , model_args.language , split='test' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) _SCREAMING_SNAKE_CASE =predict_dataset.features['label'].names # Labels _SCREAMING_SNAKE_CASE =len(_UpperCamelCase ) # Load pretrained model and tokenizer # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _SCREAMING_SNAKE_CASE =AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=_UpperCamelCase , idalabel={str(_UpperCamelCase ): label for i, label in enumerate(_UpperCamelCase )} , labelaid={label: i for i, label in enumerate(_UpperCamelCase )} , finetuning_task='xnli' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _SCREAMING_SNAKE_CASE =AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , do_lower_case=model_args.do_lower_case , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _SCREAMING_SNAKE_CASE =AutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_UpperCamelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # Preprocessing the datasets # Padding strategy if data_args.pad_to_max_length: _SCREAMING_SNAKE_CASE ='max_length' else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _SCREAMING_SNAKE_CASE =False def preprocess_function(_UpperCamelCase : int ): # Tokenize the texts return tokenizer( examples['premise'] , examples['hypothesis'] , padding=_UpperCamelCase , max_length=data_args.max_seq_length , truncation=_UpperCamelCase , ) if training_args.do_train: if data_args.max_train_samples is not None: _SCREAMING_SNAKE_CASE =min(len(_UpperCamelCase ) , data_args.max_train_samples ) _SCREAMING_SNAKE_CASE =train_dataset.select(range(_UpperCamelCase ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): _SCREAMING_SNAKE_CASE =train_dataset.map( _UpperCamelCase , batched=_UpperCamelCase , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on train dataset' , ) # Log a few random samples from the training set: for index in random.sample(range(len(_UpperCamelCase ) ) , 3 ): logger.info(f"Sample {index} of the training set: {train_dataset[index]}." ) if training_args.do_eval: if data_args.max_eval_samples is not None: _SCREAMING_SNAKE_CASE =min(len(_UpperCamelCase ) , data_args.max_eval_samples ) _SCREAMING_SNAKE_CASE =eval_dataset.select(range(_UpperCamelCase ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): _SCREAMING_SNAKE_CASE =eval_dataset.map( _UpperCamelCase , batched=_UpperCamelCase , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on validation dataset' , ) if training_args.do_predict: if data_args.max_predict_samples is not None: _SCREAMING_SNAKE_CASE =min(len(_UpperCamelCase ) , data_args.max_predict_samples ) _SCREAMING_SNAKE_CASE =predict_dataset.select(range(_UpperCamelCase ) ) with training_args.main_process_first(desc='prediction dataset map pre-processing' ): _SCREAMING_SNAKE_CASE =predict_dataset.map( _UpperCamelCase , batched=_UpperCamelCase , load_from_cache_file=not data_args.overwrite_cache , desc='Running tokenizer on prediction dataset' , ) # Get the metric function _SCREAMING_SNAKE_CASE =evaluate.load('xnli' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(_UpperCamelCase : EvalPrediction ): _SCREAMING_SNAKE_CASE =p.predictions[0] if isinstance(p.predictions , _UpperCamelCase ) else p.predictions _SCREAMING_SNAKE_CASE =np.argmax(_UpperCamelCase , axis=1 ) return metric.compute(predictions=_UpperCamelCase , references=p.label_ids ) # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _SCREAMING_SNAKE_CASE =default_data_collator elif training_args.fpaa: _SCREAMING_SNAKE_CASE =DataCollatorWithPadding(_UpperCamelCase , pad_to_multiple_of=8 ) else: _SCREAMING_SNAKE_CASE =None # Initialize our Trainer _SCREAMING_SNAKE_CASE =Trainer( model=_UpperCamelCase , args=_UpperCamelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=_UpperCamelCase , tokenizer=_UpperCamelCase , data_collator=_UpperCamelCase , ) # Training if training_args.do_train: _SCREAMING_SNAKE_CASE =None if training_args.resume_from_checkpoint is not None: _SCREAMING_SNAKE_CASE =training_args.resume_from_checkpoint elif last_checkpoint is not None: _SCREAMING_SNAKE_CASE =last_checkpoint _SCREAMING_SNAKE_CASE =trainer.train(resume_from_checkpoint=_UpperCamelCase ) _SCREAMING_SNAKE_CASE =train_result.metrics _SCREAMING_SNAKE_CASE =( data_args.max_train_samples if data_args.max_train_samples is not None else len(_UpperCamelCase ) ) _SCREAMING_SNAKE_CASE =min(_UpperCamelCase , len(_UpperCamelCase ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics('train' , _UpperCamelCase ) trainer.save_metrics('train' , _UpperCamelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) _SCREAMING_SNAKE_CASE =trainer.evaluate(eval_dataset=_UpperCamelCase ) _SCREAMING_SNAKE_CASE =data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_UpperCamelCase ) _SCREAMING_SNAKE_CASE =min(_UpperCamelCase , len(_UpperCamelCase ) ) trainer.log_metrics('eval' , _UpperCamelCase ) trainer.save_metrics('eval' , _UpperCamelCase ) # Prediction if training_args.do_predict: logger.info('*** Predict ***' ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE =trainer.predict(_UpperCamelCase , metric_key_prefix='predict' ) _SCREAMING_SNAKE_CASE =( data_args.max_predict_samples if data_args.max_predict_samples is not None else len(_UpperCamelCase ) ) _SCREAMING_SNAKE_CASE =min(_UpperCamelCase , len(_UpperCamelCase ) ) trainer.log_metrics('predict' , _UpperCamelCase ) trainer.save_metrics('predict' , _UpperCamelCase ) _SCREAMING_SNAKE_CASE =np.argmax(_UpperCamelCase , axis=1 ) _SCREAMING_SNAKE_CASE =os.path.join(training_args.output_dir , 'predictions.txt' ) if trainer.is_world_process_zero(): with open(_UpperCamelCase , 'w' ) as writer: writer.write('index\tprediction\n' ) for index, item in enumerate(_UpperCamelCase ): _SCREAMING_SNAKE_CASE =label_list[item] writer.write(f"{index}\t{item}\n" ) if __name__ == "__main__": main()
114
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase : Any = logging.get_logger(__name__) lowerCamelCase : int = { "facebook/data2vec-vision-base-ft": ( "https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json" ), } class A__ ( A__ ): A__ = 'data2vec-vision' def __init__( self : str , _a : Optional[int]=768 , _a : List[str]=12 , _a : Any=12 , _a : Tuple=3072 , _a : Optional[Any]="gelu" , _a : Tuple=0.0 , _a : str=0.0 , _a : Dict=0.02 , _a : int=1e-12 , _a : List[str]=224 , _a : str=16 , _a : Dict=3 , _a : int=False , _a : List[str]=False , _a : str=False , _a : str=False , _a : Optional[int]=0.1 , _a : List[str]=0.1 , _a : Dict=True , _a : str=[3, 5, 7, 11] , _a : List[Any]=[1, 2, 3, 6] , _a : Any=True , _a : Optional[Any]=0.4 , _a : str=256 , _a : Dict=1 , _a : List[str]=False , _a : List[Any]=255 , **_a : Any , ) -> Optional[Any]: '''simple docstring''' super().__init__(**_a ) _SCREAMING_SNAKE_CASE =hidden_size _SCREAMING_SNAKE_CASE =num_hidden_layers _SCREAMING_SNAKE_CASE =num_attention_heads _SCREAMING_SNAKE_CASE =intermediate_size _SCREAMING_SNAKE_CASE =hidden_act _SCREAMING_SNAKE_CASE =hidden_dropout_prob _SCREAMING_SNAKE_CASE =attention_probs_dropout_prob _SCREAMING_SNAKE_CASE =initializer_range _SCREAMING_SNAKE_CASE =layer_norm_eps _SCREAMING_SNAKE_CASE =image_size _SCREAMING_SNAKE_CASE =patch_size _SCREAMING_SNAKE_CASE =num_channels _SCREAMING_SNAKE_CASE =use_mask_token _SCREAMING_SNAKE_CASE =use_absolute_position_embeddings _SCREAMING_SNAKE_CASE =use_relative_position_bias _SCREAMING_SNAKE_CASE =use_shared_relative_position_bias _SCREAMING_SNAKE_CASE =layer_scale_init_value _SCREAMING_SNAKE_CASE =drop_path_rate _SCREAMING_SNAKE_CASE =use_mean_pooling # decode head attributes (semantic segmentation) _SCREAMING_SNAKE_CASE =out_indices _SCREAMING_SNAKE_CASE =pool_scales # auxiliary head attributes (semantic segmentation) _SCREAMING_SNAKE_CASE =use_auxiliary_head _SCREAMING_SNAKE_CASE =auxiliary_loss_weight _SCREAMING_SNAKE_CASE =auxiliary_channels _SCREAMING_SNAKE_CASE =auxiliary_num_convs _SCREAMING_SNAKE_CASE =auxiliary_concat_input _SCREAMING_SNAKE_CASE =semantic_loss_ignore_index class A__ ( A__ ): A__ = version.parse('1.11' ) @property def A ( self : str ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def A ( self : List[Any] ) -> float: '''simple docstring''' return 1e-4
114
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase = { 'configuration_timesformer': ['TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimesformerConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ 'TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimesformerModel', 'TimesformerForVideoClassification', 'TimesformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
29
def lowerCamelCase_ ( _a : List[str] ): '''simple docstring''' UpperCAmelCase_ : Tuple = [0] * len(_a ) UpperCAmelCase_ : Dict = [] UpperCAmelCase_ : Optional[int] = [] UpperCAmelCase_ : Dict = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(_a ) ): if indegree[i] == 0: queue.append(_a ) while queue: UpperCAmelCase_ : List[str] = queue.pop(0 ) cnt += 1 topo.append(_a ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(_a ) if cnt != len(_a ): print("""Cycle exists""" ) else: print(_a ) # Adjacency List of Graph UpperCamelCase_ = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
345
0
from __future__ import annotations from collections.abc import Sequence from typing import Literal def snake_case (__lowercase , __lowercase ) -> str | Literal[False]: '''simple docstring''' _snake_case : Dict = list(__lowercase ) _snake_case : Union[str, Any] = list(__lowercase ) _snake_case : Optional[int] = 0 for i in range(len(__lowercase ) ): if lista[i] != lista[i]: count += 1 _snake_case : Tuple = "_" if count > 1: return False else: return "".join(__lowercase ) def snake_case (__lowercase ) -> list[str]: '''simple docstring''' _snake_case : Tuple = [] while True: _snake_case : Any = ["$"] * len(__lowercase ) _snake_case : Dict = [] for i in range(len(__lowercase ) ): for j in range(i + 1 , len(__lowercase ) ): _snake_case : Optional[Any] = compare_string(binary[i] , binary[j] ) if k is False: _snake_case : int = "*" _snake_case : int = "*" temp.append("X" ) for i in range(len(__lowercase ) ): if checka[i] == "$": pi.append(binary[i] ) if len(__lowercase ) == 0: return pi _snake_case : List[Any] = list(set(__lowercase ) ) def snake_case (__lowercase , __lowercase ) -> list[str]: '''simple docstring''' _snake_case : Optional[int] = [] for minterm in minterms: _snake_case : Union[str, Any] = "" for _ in range(__lowercase ): _snake_case : Dict = str(minterm % 2 ) + string minterm //= 2 temp.append(__lowercase ) return temp def snake_case (__lowercase , __lowercase , __lowercase ) -> bool: '''simple docstring''' _snake_case : str = list(__lowercase ) _snake_case : Tuple = list(__lowercase ) _snake_case : List[str] = 0 for i in range(len(__lowercase ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def snake_case (__lowercase , __lowercase ) -> list[str]: '''simple docstring''' _snake_case : int = [] _snake_case : Dict = [0] * len(__lowercase ) for i in range(len(chart[0] ) ): _snake_case : Any = 0 _snake_case : str = -1 for j in range(len(__lowercase ) ): if chart[j][i] == 1: count += 1 _snake_case : Tuple = j if count == 1: _snake_case : Optional[Any] = 1 for i in range(len(__lowercase ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(__lowercase ) ): _snake_case : Optional[int] = 0 temp.append(prime_implicants[i] ) while True: _snake_case : int = 0 _snake_case : Any = -1 _snake_case : Optional[int] = 0 for i in range(len(__lowercase ) ): _snake_case : Optional[Any] = chart[i].count(1 ) if count_n > max_n: _snake_case : Optional[int] = count_n _snake_case : int = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(__lowercase ) ): _snake_case : Optional[int] = 0 def snake_case (__lowercase , __lowercase ) -> list[list[int]]: '''simple docstring''' _snake_case : Dict = [[0 for x in range(len(__lowercase ) )] for x in range(len(__lowercase ) )] for i in range(len(__lowercase ) ): _snake_case : List[str] = prime_implicants[i].count("_" ) for j in range(len(__lowercase ) ): if is_for_table(prime_implicants[i] , binary[j] , __lowercase ): _snake_case : List[str] = 1 return chart def snake_case () -> None: '''simple docstring''' _snake_case : Optional[int] = int(input("Enter the no. of variables\n" ) ) _snake_case : int = [ float(__lowercase ) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n" ).split() ] _snake_case : Dict = decimal_to_binary(__lowercase , __lowercase ) _snake_case : Any = check(__lowercase ) print("Prime Implicants are:" ) print(__lowercase ) _snake_case : Dict = prime_implicant_chart(__lowercase , __lowercase ) _snake_case : Optional[int] = selection(__lowercase , __lowercase ) print("Essential Prime Implicants are:" ) print(__lowercase ) if __name__ == "__main__": import doctest doctest.testmod() main()
284
def snake_case (__lowercase ) -> bool: '''simple docstring''' _snake_case : Dict = 0 for ch in input_str: _snake_case : int = ord(__lowercase ) _snake_case : List[Any] = pow(2 , __lowercase ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
284
1
'''simple docstring''' from __future__ import annotations from typing import Generic, TypeVar A_ : Union[str, Any] = TypeVar("""T""") class lowercase ( Generic[T] ): """simple docstring""" def __init__( self ,a_ ) -> None: _UpperCAmelCase : str = data _UpperCAmelCase : List[str] = self _UpperCAmelCase : Optional[Any] = 0 class lowercase ( Generic[T] ): """simple docstring""" def __init__( self ) -> None: # map from node name to the node object _UpperCAmelCase : dict[T, DisjointSetTreeNode[T]] = {} def _snake_case ( self ,a_ ) -> None: # create a new set with x as its member _UpperCAmelCase : Tuple = DisjointSetTreeNode(a_ ) def _snake_case ( self ,a_ ) -> DisjointSetTreeNode[T]: # find the set x belongs to (with path-compression) _UpperCAmelCase : str = self.map[data] if elem_ref != elem_ref.parent: _UpperCAmelCase : Tuple = self.find_set(elem_ref.parent.data ) return elem_ref.parent def _snake_case ( self ,a_ ,a_ ) -> None: # helper function for union operation if nodea.rank > nodea.rank: _UpperCAmelCase : str = nodea else: _UpperCAmelCase : str = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def _snake_case ( self ,a_ ,a_ ) -> None: # merge 2 disjoint sets self.link(self.find_set(a_ ) ,self.find_set(a_ ) ) class lowercase ( Generic[T] ): """simple docstring""" def __init__( self ) -> None: # connections: map from the node to the neighbouring nodes (with weights) _UpperCAmelCase : dict[T, dict[T, int]] = {} def _snake_case ( self ,a_ ) -> None: # add a node ONLY if its not present in the graph if node not in self.connections: _UpperCAmelCase : str = {} def _snake_case ( self ,a_ ,a_ ,a_ ) -> None: # add an edge with the given weight self.add_node(a_ ) self.add_node(a_ ) _UpperCAmelCase : Optional[int] = weight _UpperCAmelCase : str = weight def _snake_case ( self ) -> GraphUndirectedWeighted[T]: _UpperCAmelCase : List[Any] = [] _UpperCAmelCase : Tuple = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda a_ : x[2] ) # creating the disjoint set _UpperCAmelCase : List[str] = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(a_ ) # MST generation _UpperCAmelCase : Tuple = 0 _UpperCAmelCase : List[Any] = 0 _UpperCAmelCase : List[str] = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase : Any = edges[index] index += 1 _UpperCAmelCase : Optional[int] = disjoint_set.find_set(a_ ) _UpperCAmelCase : List[str] = disjoint_set.find_set(a_ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(a_ ,a_ ,a_ ) disjoint_set.union(a_ ,a_ ) return graph
215
'''simple docstring''' from __future__ import annotations import typing from collections import Counter def snake_case_ ( lowerCAmelCase_ )-> typing.Counter[int]: '''simple docstring''' _UpperCAmelCase : typing.Counter[int] = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(lowerCAmelCase_ , max_perimeter + 1 ): _UpperCAmelCase : List[str] = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(lowerCAmelCase_ ): _UpperCAmelCase : Optional[Any] = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def snake_case_ ( lowerCAmelCase_ = 1000 )-> int: '''simple docstring''' _UpperCAmelCase : int = pythagorean_triple(lowerCAmelCase_ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(f"""Perimeter {solution()} has maximum solutions""")
215
1
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase (a_ :str , a_ :Union[str, Any] , a_ :str , a_ :Any , a_ :List[str]) -> Union[str, Any]: # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file lowercase :List[str] = TapasConfig.from_json_file(a_) # set absolute/relative position embeddings parameter lowercase :Any = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": lowercase :Union[str, Any] = TapasForQuestionAnswering(config=a_) elif task == "WTQ": # run_task_main.py hparams lowercase :Any = 4 lowercase :Union[str, Any] = True # hparam_utils.py hparams lowercase :Any = 0.66_46_94 lowercase :Optional[int] = 0.20_79_51 lowercase :Tuple = 0.12_11_94 lowercase :List[Any] = True lowercase :int = True lowercase :Optional[Any] = False lowercase :List[str] = 0.0_35_25_13 lowercase :Union[str, Any] = TapasForQuestionAnswering(config=a_) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams lowercase :List[Any] = 4 lowercase :List[str] = False # hparam_utils.py hparams lowercase :int = 36.45_19 lowercase :str = 0.90_34_21 lowercase :Union[str, Any] = 2_22.0_88 lowercase :Optional[int] = True lowercase :Dict = True lowercase :str = True lowercase :Dict = 0.76_31_41 lowercase :int = TapasForQuestionAnswering(config=a_) elif task == "TABFACT": lowercase :Dict = TapasForSequenceClassification(config=a_) elif task == "MLM": lowercase :Optional[Any] = TapasForMaskedLM(config=a_) elif task == "INTERMEDIATE_PRETRAINING": lowercase :List[str] = TapasModel(config=a_) else: raise ValueError(F"""Task {task} not supported.""") print(F"""Building PyTorch model from configuration: {config}""") # Load weights from tf checkpoint load_tf_weights_in_tapas(a_ , a_ , a_) # Save pytorch-model (weights and configuration) print(F"""Save PyTorch model to {pytorch_dump_path}""") model.save_pretrained(a_) # Save tokenizer files print(F"""Save tokenizer files to {pytorch_dump_path}""") lowercase :List[str] = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + '''vocab.txt''' , model_max_length=512) tokenizer.save_pretrained(a_) print('''Used relative position embeddings:''' , model.config.reset_position_index_per_cell) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''SQA''', type=str, help='''Model task for which to convert a checkpoint. Defaults to SQA.''' ) parser.add_argument( '''--reset_position_index_per_cell''', default=False, action='''store_true''', help='''Whether to use relative position embeddings or not. Defaults to True.''', ) parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--tapas_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained TAPAS model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) UpperCAmelCase = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
172
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class __magic_name__ ( __UpperCAmelCase ): __A : UNetaDModel __A : ScoreSdeVeScheduler def __init__( self : List[Any] , snake_case__ : UNetaDModel , snake_case__ : ScoreSdeVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=snake_case__ , scheduler=snake_case__ ) @torch.no_grad() def __call__( self : Optional[Any] , snake_case__ : int = 1 , snake_case__ : int = 2_0_0_0 , snake_case__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , snake_case__ : Optional[str] = "pil" , snake_case__ : bool = True , **snake_case__ : Optional[Any] , ): '''simple docstring''' lowercase :List[str] = self.unet.config.sample_size lowercase :Optional[int] = (batch_size, 3, img_size, img_size) lowercase :Optional[Any] = self.unet lowercase :Dict = randn_tensor(snake_case__ , generator=snake_case__ ) * self.scheduler.init_noise_sigma lowercase :Any = sample.to(self.device ) self.scheduler.set_timesteps(snake_case__ ) self.scheduler.set_sigmas(snake_case__ ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): lowercase :Union[str, Any] = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): lowercase :Union[str, Any] = self.unet(snake_case__ , snake_case__ ).sample lowercase :Union[str, Any] = self.scheduler.step_correct(snake_case__ , snake_case__ , generator=snake_case__ ).prev_sample # prediction step lowercase :List[str] = model(snake_case__ , snake_case__ ).sample lowercase :Optional[int] = self.scheduler.step_pred(snake_case__ , snake_case__ , snake_case__ , generator=snake_case__ ) lowercase , lowercase :Optional[int] = output.prev_sample, output.prev_sample_mean lowercase :List[Any] = sample_mean.clamp(0 , 1 ) lowercase :Optional[Any] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowercase :Dict = self.numpy_to_pil(snake_case__ ) if not return_dict: return (sample,) return ImagePipelineOutput(images=snake_case__ )
172
1
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor A : str = logging.get_logger(__name__) class __A( A__ ): def __init__( self , *_snake_case , **_snake_case ) -> None: '''simple docstring''' warnings.warn( '''The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use BeitImageProcessor instead.''' , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
6
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def A ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' with open(snake_case__ ) as metadata_file: SCREAMING_SNAKE_CASE__ = json.load(snake_case__ ) SCREAMING_SNAKE_CASE__ = LukeConfig(use_entity_aware_attention=snake_case__ , **metadata["""model_config"""] ) # Load in the weights from the checkpoint_path SCREAMING_SNAKE_CASE__ = torch.load(snake_case__ , map_location="""cpu""" ) # Load the entity vocab file SCREAMING_SNAKE_CASE__ = load_entity_vocab(snake_case__ ) SCREAMING_SNAKE_CASE__ = RobertaTokenizer.from_pretrained(metadata["""model_config"""]["""bert_model_name"""] ) # Add special tokens to the token vocabulary for downstream tasks SCREAMING_SNAKE_CASE__ = AddedToken("""<ent>""" , lstrip=snake_case__ , rstrip=snake_case__ ) SCREAMING_SNAKE_CASE__ = AddedToken("""<ent2>""" , lstrip=snake_case__ , rstrip=snake_case__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(f"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(snake_case__ ) with open(os.path.join(snake_case__ , LukeTokenizer.vocab_files_names["""entity_vocab_file"""] ) , """w""" ) as f: json.dump(snake_case__ , snake_case__ ) SCREAMING_SNAKE_CASE__ = LukeTokenizer.from_pretrained(snake_case__ ) # Initialize the embeddings of the special tokens SCREAMING_SNAKE_CASE__ = state_dict["""embeddings.word_embeddings.weight"""] SCREAMING_SNAKE_CASE__ = word_emb[tokenizer.convert_tokens_to_ids(["""@"""] )[0]].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = word_emb[tokenizer.convert_tokens_to_ids(["""#"""] )[0]].unsqueeze(0 ) SCREAMING_SNAKE_CASE__ = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: SCREAMING_SNAKE_CASE__ = f"""encoder.layer.{layer_index}.attention.self.""" SCREAMING_SNAKE_CASE__ = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE__ = state_dict[prefix + matrix_name] SCREAMING_SNAKE_CASE__ = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks SCREAMING_SNAKE_CASE__ = state_dict["""entity_embeddings.entity_embeddings.weight"""] SCREAMING_SNAKE_CASE__ = entity_emb[entity_vocab["""[MASK]"""]] SCREAMING_SNAKE_CASE__ = LukeModel(config=snake_case__ ).eval() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = model.load_state_dict(snake_case__ , strict=snake_case__ ) if not (len(snake_case__ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(f"""Missing keys {", ".join(snake_case__ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith("""entity_predictions""" ) or key.startswith("""lm_head""" ) for key in unexpected_keys )): raise ValueError( """Unexpected keys""" f""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs SCREAMING_SNAKE_CASE__ = LukeTokenizer.from_pretrained(snake_case__ , task="""entity_classification""" ) SCREAMING_SNAKE_CASE__ = ( """Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the""" """ new world number one avoid a humiliating second- round exit at Wimbledon .""" ) SCREAMING_SNAKE_CASE__ = (39, 42) SCREAMING_SNAKE_CASE__ = tokenizer(snake_case__ , entity_spans=[span] , add_prefix_space=snake_case__ , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = model(**snake_case__ ) # Verify word hidden states if model_size == "large": SCREAMING_SNAKE_CASE__ = torch.Size((1, 42, 10_24) ) SCREAMING_SNAKE_CASE__ = torch.tensor( [[0.01_33, 0.08_65, 0.00_95], [0.30_93, -0.25_76, -0.74_18], [-0.17_20, -0.21_17, -0.28_69]] ) else: # base SCREAMING_SNAKE_CASE__ = torch.Size((1, 42, 7_68) ) SCREAMING_SNAKE_CASE__ = torch.tensor([[0.00_37, 0.13_68, -0.00_91], [0.10_99, 0.33_29, -0.10_95], [0.07_65, 0.53_35, 0.11_79]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": SCREAMING_SNAKE_CASE__ = torch.Size((1, 1, 10_24) ) SCREAMING_SNAKE_CASE__ = torch.tensor([[0.04_66, -0.01_06, -0.01_79]] ) else: # base SCREAMING_SNAKE_CASE__ = torch.Size((1, 1, 7_68) ) SCREAMING_SNAKE_CASE__ = torch.tensor([[0.14_57, 0.10_44, 0.01_74]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( f"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" f""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , snake_case__ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print("""Saving PyTorch model to {}""".format(snake_case__ ) ) model.save_pretrained(snake_case__ ) def A ( snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = {} with open(snake_case__ , """r""" , encoding="""utf-8""" ) as f: for index, line in enumerate(snake_case__ ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = line.rstrip().split("""\t""" ) SCREAMING_SNAKE_CASE__ = index return entity_vocab if __name__ == "__main__": A_ : str = argparse.ArgumentParser() # Required parameters parser.add_argument("--checkpoint_path", type=str, help="Path to a pytorch_model.bin file.") parser.add_argument( "--metadata_path", default=None, type=str, help="Path to a metadata.json file, defining the configuration." ) parser.add_argument( "--entity_vocab_path", default=None, type=str, help="Path to an entity_vocab.tsv file, containing the entity vocabulary.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to where to dump the output PyTorch model." ) parser.add_argument( "--model_size", default="base", type=str, choices=["base", "large"], help="Size of the model to be converted." ) A_ : Optional[Any] = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
165
0
import argparse import json import subprocess def UpperCamelCase ( __lowercase : Optional[int] ,__lowercase : Tuple ): '''simple docstring''' A_ : List[str] = [] A_ : Optional[int] = ( f'''curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"''' ' https://api.github.com/repos/huggingface/transformers/actions/runners' ) A_ : Union[str, Any] = subprocess.run(__lowercase ,shell=__lowercase ,stdout=subprocess.PIPE ) A_ : Optional[Any] = output.stdout.decode('utf-8' ) A_ : Tuple = json.loads(__lowercase ) A_ : Tuple = status['runners'] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__lowercase ) # save the result so we can report them on Slack with open('offline_runners.txt' ,'w' ) as fp: fp.write(json.dumps(__lowercase ) ) if len(__lowercase ) > 0: A_ : Dict = '\n'.join([x['name'] for x in offline_runners] ) raise ValueError(f'''The following runners are offline:\n{failed}''' ) if __name__ == "__main__": def UpperCamelCase ( __lowercase : Union[str, Any] ): '''simple docstring''' return values.split(',' ) _UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--target_runners""", default=None, type=list_str, required=True, help="""Comma-separated list of runners to check status.""", ) parser.add_argument( """--token""", default=None, type=str, required=True, help="""A token that has actions:read permission.""" ) _UpperCAmelCase = parser.parse_args() get_runner_status(args.target_runners, args.token)
371
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { """SCUT-DLVCLab/lilt-roberta-en-base""": ( """https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json""" ), } class UpperCAmelCase ( __A ): '''simple docstring''' lowerCamelCase_ = '''lilt''' def __init__( self , lowercase=3_0_5_2_2 , lowercase=7_6_8 , lowercase=1_2 , lowercase=1_2 , lowercase=3_0_7_2 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_1_2 , lowercase=2 , lowercase=0.02 , lowercase=1E-12 , lowercase=0 , lowercase="absolute" , lowercase=None , lowercase=4 , lowercase=1_0_2_4 , **lowercase , ): """simple docstring""" super().__init__(pad_token_id=lowercase , **lowercase ) A_ : str = vocab_size A_ : Tuple = hidden_size A_ : List[Any] = num_hidden_layers A_ : Any = num_attention_heads A_ : Union[str, Any] = hidden_act A_ : int = intermediate_size A_ : int = hidden_dropout_prob A_ : Union[str, Any] = attention_probs_dropout_prob A_ : List[str] = max_position_embeddings A_ : List[str] = type_vocab_size A_ : Any = initializer_range A_ : str = layer_norm_eps A_ : Optional[int] = position_embedding_type A_ : Union[str, Any] = classifier_dropout A_ : Optional[Any] = channel_shrink_ratio A_ : int = max_ad_position_embeddings
192
0
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger lowerCAmelCase__ :Union[str, Any] = get_logger(__name__) lowerCAmelCase__ :List[str] = R''' Args: input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam search or log softmax for each vocabulary token when using beam search kwargs (`Dict[str, Any]`, *optional*): Additional logits processor specific kwargs. Return: `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores. ''' class __a : @add_start_docstrings(_SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class __a : @add_start_docstrings(_SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" raise NotImplementedError( f'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class __a ( UpperCAmelCase ): @add_start_docstrings(_SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" for processor in self: _UpperCAmelCase = inspect.signature(processor.__call__ ).parameters if len(_SCREAMING_SNAKE_CASE ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( f'''Make sure that all the required parameters: {list(function_args.keys() )} for ''' f'''{processor.__class__} are passed to the logits processor.''' ) _UpperCAmelCase = processor(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) else: _UpperCAmelCase = processor(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or not (temperature > 0): raise ValueError(f'''`temperature` has to be a strictly positive float, but is {temperature}''' ) _UpperCAmelCase = temperature def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase = scores / self.temperature return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = -float('Inf' ) , _SCREAMING_SNAKE_CASE = 1 ) -> str: """simple docstring""" if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or (top_p < 0 or top_p > 1.0): raise ValueError(f'''`top_p` has to be a float > 0 and < 1, but is {top_p}''' ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or (min_tokens_to_keep < 1): raise ValueError(f'''`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}''' ) _UpperCAmelCase = top_p _UpperCAmelCase = filter_value _UpperCAmelCase = min_tokens_to_keep def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = lax.top_k(_SCREAMING_SNAKE_CASE , scores.shape[-1] ) _UpperCAmelCase = jnp.full_like(_SCREAMING_SNAKE_CASE , self.filter_value ) _UpperCAmelCase = jax.nn.softmax(_SCREAMING_SNAKE_CASE , axis=-1 ).cumsum(axis=-1 ) _UpperCAmelCase = cumulative_probs < self.top_p # include the token that is higher than top_p as well _UpperCAmelCase = jnp.roll(_SCREAMING_SNAKE_CASE , 1 ) score_mask |= score_mask.at[:, 0].set(_SCREAMING_SNAKE_CASE ) # min tokens to keep _UpperCAmelCase = score_mask.at[:, : self.min_tokens_to_keep].set(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.where(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jax.lax.sort_key_val(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )[-1] return next_scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = -float('Inf' ) , _SCREAMING_SNAKE_CASE = 1 ) -> Optional[int]: """simple docstring""" if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or top_k <= 0: raise ValueError(f'''`top_k` has to be a strictly positive integer, but is {top_k}''' ) _UpperCAmelCase = max(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = filter_value def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = scores.shape _UpperCAmelCase = jnp.full(batch_size * vocab_size , self.filter_value ) _UpperCAmelCase = min(self.top_k , scores.shape[-1] ) # Safety check _UpperCAmelCase , _UpperCAmelCase = lax.top_k(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.broadcast_to((jnp.arange(_SCREAMING_SNAKE_CASE ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() _UpperCAmelCase = topk_scores.flatten() _UpperCAmelCase = topk_indices.flatten() + shift _UpperCAmelCase = next_scores_flat.at[topk_indices_flat].set(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = next_scores_flat.reshape(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return next_scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" _UpperCAmelCase = bos_token_id def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase = jnp.full(scores.shape , -float('inf' ) ) _UpperCAmelCase = 1 - jnp.bool_(cur_len - 1 ) _UpperCAmelCase = jnp.where(_SCREAMING_SNAKE_CASE , new_scores.at[:, self.bos_token_id].set(0 ) , _SCREAMING_SNAKE_CASE ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = max_length _UpperCAmelCase = eos_token_id def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase = jnp.full(scores.shape , -float('inf' ) ) _UpperCAmelCase = 1 - jnp.bool_(cur_len - self.max_length + 1 ) _UpperCAmelCase = jnp.where(_SCREAMING_SNAKE_CASE , new_scores.at[:, self.eos_token_id].set(0 ) , _SCREAMING_SNAKE_CASE ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or min_length < 0: raise ValueError(f'''`min_length` has to be a positive integer, but is {min_length}''' ) if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or eos_token_id < 0: raise ValueError(f'''`eos_token_id` has to be a positive integer, but is {eos_token_id}''' ) _UpperCAmelCase = min_length _UpperCAmelCase = eos_token_id def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) _UpperCAmelCase = jnp.where(_SCREAMING_SNAKE_CASE , scores.at[:, self.eos_token_id].set(-float('inf' ) ) , _SCREAMING_SNAKE_CASE ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" _UpperCAmelCase = list(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = begin_index def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" _UpperCAmelCase = 1 - jnp.bool_(cur_len - self.begin_index ) _UpperCAmelCase = jnp.where(_SCREAMING_SNAKE_CASE , scores.at[:, self.begin_suppress_tokens].set(-float('inf' ) ) , _SCREAMING_SNAKE_CASE ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" _UpperCAmelCase = list(_SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" _UpperCAmelCase = scores.at[..., self.suppress_tokens].set(-float('inf' ) ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = dict(_SCREAMING_SNAKE_CASE ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. _UpperCAmelCase = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: _UpperCAmelCase = force_token_array.at[index].set(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.intaa(_SCREAMING_SNAKE_CASE ) def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> jnp.ndarray: """simple docstring""" def _force_token(_SCREAMING_SNAKE_CASE ): _UpperCAmelCase = scores.shape[0] _UpperCAmelCase = self.force_token_array[generation_idx] _UpperCAmelCase = jnp.ones_like(_SCREAMING_SNAKE_CASE , dtype=scores.dtype ) * -float('inf' ) _UpperCAmelCase = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) _UpperCAmelCase = lax.dynamic_update_slice(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , (0, current_token) ) return new_scores _UpperCAmelCase = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(_SCREAMING_SNAKE_CASE ) , lambda: scores , ) , ) return scores class __a ( UpperCAmelCase ): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" _UpperCAmelCase = generate_config.eos_token_id _UpperCAmelCase = generate_config.no_timestamps_token_id _UpperCAmelCase = generate_config.no_timestamps_token_id + 1 _UpperCAmelCase = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(_SCREAMING_SNAKE_CASE , 'max_initial_timestamp_index' ): _UpperCAmelCase = generate_config.max_initial_timestamp_index else: _UpperCAmelCase = model_config.vocab_size if self.max_initial_timestamp_index is None: _UpperCAmelCase = model_config.vocab_size def __call__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" _UpperCAmelCase = scores.at[:, self.no_timestamps_token_id].set(-float('inf' ) ) def handle_pairs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _UpperCAmelCase = jnp.where((cur_len - self.begin_index) >= 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , _SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = jnp.where((cur_len - self.begin_index) < 2 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) return jnp.where( _SCREAMING_SNAKE_CASE , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float('inf' ) ) , scores_k.at[: self.eos_token_id].set(-float('inf' ) ) , ) , _SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = jax.vmap(_SCREAMING_SNAKE_CASE )(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.where(cur_len == self.begin_index , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , _SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = self.timestamp_begin + self.max_initial_timestamp_index _UpperCAmelCase = jnp.where( _SCREAMING_SNAKE_CASE , scores.at[:, last_allowed + 1 :].set(-float('inf' ) ) , _SCREAMING_SNAKE_CASE , ) # if sum of probability over timestamps is above any other token, sample timestamp _UpperCAmelCase = jax.nn.log_softmax(_SCREAMING_SNAKE_CASE , axis=-1 ) def handle_cumulative_probs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): _UpperCAmelCase = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) _UpperCAmelCase = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float('inf' ) ) , _SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = jax.vmap(_SCREAMING_SNAKE_CASE )(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return scores
329
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def lowerCAmelCase__ ( *a__: str , a__: Optional[Union[Dict, Any]] = None , a__: Dict=True , a__: Any=2 ) -> Union[str, Any]: '''simple docstring''' from .. import __version__ _UpperCAmelCase = take_from _UpperCAmelCase = () if not isinstance(args[0] , a__ ): _UpperCAmelCase = (args,) for attribute, version_name, message in args: if version.parse(version.parse(a__ ).base_version ) >= version.parse(a__ ): raise ValueError( F'''The deprecation tuple {(attribute, version_name, message)} should be removed since diffusers\'''' F''' version {__version__} is >= {version_name}''' ) _UpperCAmelCase = None if isinstance(a__ , a__ ) and attribute in deprecated_kwargs: values += (deprecated_kwargs.pop(a__ ),) _UpperCAmelCase = F'''The `{attribute}` argument is deprecated and will be removed in version {version_name}.''' elif hasattr(a__ , a__ ): values += (getattr(a__ , a__ ),) _UpperCAmelCase = F'''The `{attribute}` attribute is deprecated and will be removed in version {version_name}.''' elif deprecated_kwargs is None: _UpperCAmelCase = F'''`{attribute}` is deprecated and will be removed in version {version_name}.''' if warning is not None: _UpperCAmelCase = warning + ' ' if standard_warn else '' warnings.warn(warning + message , a__ , stacklevel=a__ ) if isinstance(a__ , a__ ) and len(a__ ) > 0: _UpperCAmelCase = inspect.getouterframes(inspect.currentframe() )[1] _UpperCAmelCase = call_frame.filename _UpperCAmelCase = call_frame.lineno _UpperCAmelCase = call_frame.function _UpperCAmelCase , _UpperCAmelCase = next(iter(deprecated_kwargs.items() ) ) raise TypeError(F'''{function} in {filename} line {line_number-1} got an unexpected keyword argument `{key}`''' ) if len(a__ ) == 0: return elif len(a__ ) == 1: return values[0] return values
329
1
'''simple docstring''' import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def snake_case__ ( ) -> int: with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(A__ ): requests.request('''GET''' , '''https://huggingface.co''' ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request('''GET''' , '''https://huggingface.co''' , timeout=1.0 ) @pytest.mark.integration def snake_case__ ( ) -> str: with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request('''GET''' , '''https://huggingface.co''' ) def snake_case__ ( ) -> str: with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(A__ ): http_head('''https://huggingface.co''' )
357
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import AlignProcessor, EfficientNetImageProcessor @require_vision class UpperCamelCase_ (unittest.TestCase ): """simple docstring""" def _a ( self : Union[str, Any] ): """simple docstring""" A_ : Any = tempfile.mkdtemp() A_ : List[Any] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest''', ] A_ : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) A_ : Tuple = { '''do_resize''': True, '''size''': 20, '''do_center_crop''': True, '''crop_size''': 18, '''do_normalize''': True, '''image_mean''': [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73], '''image_std''': [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11], } A_ : List[Any] = os.path.join(self.tmpdirname , _lowerCamelCase ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(_lowerCamelCase , _lowerCamelCase ) def _a ( self : Dict , **_lowerCamelCase : Tuple ): """simple docstring""" return BertTokenizer.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def _a ( self : Optional[int] , **_lowerCamelCase : Optional[int] ): """simple docstring""" return BertTokenizerFast.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def _a ( self : Optional[Any] , **_lowerCamelCase : Tuple ): """simple docstring""" return EfficientNetImageProcessor.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def _a ( self : Tuple ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def _a ( self : int ): """simple docstring""" A_ : Union[str, Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] A_ : Any = [Image.fromarray(np.moveaxis(_lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _a ( self : int ): """simple docstring""" A_ : Tuple = self.get_tokenizer() A_ : Tuple = self.get_rust_tokenizer() A_ : Dict = self.get_image_processor() A_ : List[Any] = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) A_ : str = AlignProcessor.from_pretrained(self.tmpdirname , use_fast=_lowerCamelCase ) A_ : Any = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) A_ : List[Any] = AlignProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , _lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , _lowerCamelCase ) def _a ( self : List[Any] ): """simple docstring""" A_ : List[str] = AlignProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) A_ : Optional[int] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) A_ : Tuple = self.get_image_processor(do_normalize=_lowerCamelCase , padding_value=1.0 ) A_ : List[str] = AlignProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=_lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _lowerCamelCase ) def _a ( self : Union[str, Any] ): """simple docstring""" A_ : Dict = self.get_image_processor() A_ : Any = self.get_tokenizer() A_ : List[str] = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : Any = self.prepare_image_inputs() A_ : List[Any] = image_processor(_lowerCamelCase , return_tensors='''np''' ) A_ : str = processor(images=_lowerCamelCase , return_tensors='''np''' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _a ( self : Dict ): """simple docstring""" A_ : str = self.get_image_processor() A_ : List[str] = self.get_tokenizer() A_ : Optional[int] = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : int = '''lower newer''' A_ : str = processor(text=_lowerCamelCase ) A_ : Dict = tokenizer(_lowerCamelCase , padding='''max_length''' , max_length=64 ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _a ( self : str ): """simple docstring""" A_ : Optional[int] = self.get_image_processor() A_ : Optional[Any] = self.get_tokenizer() A_ : List[str] = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : List[Any] = '''lower newer''' A_ : Optional[int] = self.prepare_image_inputs() A_ : List[Any] = processor(text=_lowerCamelCase , images=_lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(_lowerCamelCase ): processor() def _a ( self : List[str] ): """simple docstring""" A_ : Optional[Any] = self.get_image_processor() A_ : Optional[int] = self.get_tokenizer() A_ : List[Any] = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : str = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] A_ : str = processor.batch_decode(_lowerCamelCase ) A_ : Union[str, Any] = tokenizer.batch_decode(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def _a ( self : Tuple ): """simple docstring""" A_ : str = self.get_image_processor() A_ : Tuple = self.get_tokenizer() A_ : Any = AlignProcessor(tokenizer=_lowerCamelCase , image_processor=_lowerCamelCase ) A_ : str = '''lower newer''' A_ : List[str] = self.prepare_image_inputs() A_ : Tuple = processor(text=_lowerCamelCase , images=_lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
4
0
from __future__ import annotations lowerCamelCase : List[Any] = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,) -> Union[str, Any]: snake_case : Union[str, Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__lowerCamelCase ) ) ] # the reference grid snake_case : List[Any] = 1 snake_case : Optional[int] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__lowerCamelCase ) ) ] # the action grid snake_case : Dict = init[0] snake_case : int = init[1] snake_case : Union[str, Any] = 0 snake_case : Union[str, Any] = g + heuristic[x][y] # cost from starting cell to destination cell snake_case : Dict = [[f, g, x, y]] snake_case : Optional[int] = False # flag that is set when search is complete snake_case : Optional[Any] = False # flag set if we can't find expand while not found and not resign: if len(__lowerCamelCase ) == 0: raise ValueError("""Algorithm is unable to find solution""" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() snake_case : int = cell.pop() snake_case : Tuple = next_cell[2] snake_case : Any = next_cell[3] snake_case : Tuple = next_cell[1] if x == goal[0] and y == goal[1]: snake_case : List[Any] = True else: for i in range(len(__lowerCamelCase ) ): # to try out different valid actions snake_case : Union[str, Any] = x + DIRECTIONS[i][0] snake_case : Union[str, Any] = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__lowerCamelCase ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: snake_case : List[str] = g + cost snake_case : Optional[Any] = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) snake_case : Dict = 1 snake_case : Dict = i snake_case : Union[str, Any] = [] snake_case : Tuple = goal[0] snake_case : List[str] = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: snake_case : Tuple = x - DIRECTIONS[action[x][y]][0] snake_case : Any = y - DIRECTIONS[action[x][y]][1] snake_case : Tuple = xa snake_case : int = ya invpath.append([x, y] ) snake_case : List[Any] = [] for i in range(len(__lowerCamelCase ) ): path.append(invpath[len(__lowerCamelCase ) - 1 - i] ) return path, action if __name__ == "__main__": lowerCamelCase : Dict = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] lowerCamelCase : List[Any] = [0, 0] # all coordinates are given in format [y,x] lowerCamelCase : Union[str, Any] = [len(grid) - 1, len(grid[0]) - 1] lowerCamelCase : Any = 1 # the cost map which pushes the path closer to the goal lowerCamelCase : List[str] = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): lowerCamelCase : List[str] = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map lowerCamelCase : Dict = 9_9 lowerCamelCase , lowerCamelCase : Optional[Any] = search(grid, init, goal, cost, heuristic) print('ACTION MAP') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
124
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): __UpperCAmelCase = "pt" elif is_tf_available(): __UpperCAmelCase = "tf" else: __UpperCAmelCase = "jax" class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ =ByTaTokenizer UpperCAmelCase_ =False def _UpperCamelCase ( self ) -> Tuple: super().setUp() SCREAMING_SNAKE_CASE_ = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _UpperCamelCase ( self ) -> List[str]: return ByTaTokenizer.from_pretrained('''google/byt5-small''' ) def _UpperCamelCase ( self , **_A ) -> ByTaTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **_A ) def _UpperCamelCase ( self , _A , _A=False , _A=20 , _A=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. SCREAMING_SNAKE_CASE_ = [] for i in range(len(_A ) ): try: SCREAMING_SNAKE_CASE_ = tokenizer.decode([i] , clean_up_tokenization_spaces=_A ) except UnicodeDecodeError: pass toks.append((i, tok) ) SCREAMING_SNAKE_CASE_ = list(filter(lambda _A : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , _A ) ) SCREAMING_SNAKE_CASE_ = list(filter(lambda _A : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_A ) , _A ) ) if max_length is not None and len(_A ) > max_length: SCREAMING_SNAKE_CASE_ = toks[:max_length] if min_length is not None and len(_A ) < min_length and len(_A ) > 0: while len(_A ) < min_length: SCREAMING_SNAKE_CASE_ = toks + toks # toks_str = [t[1] for t in toks] SCREAMING_SNAKE_CASE_ = [t[0] for t in toks] # Ensure consistency SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A , clean_up_tokenization_spaces=_A ) if " " not in output_txt and len(_A ) > 1: SCREAMING_SNAKE_CASE_ = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_A ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_A ) ) if with_prefix_space: SCREAMING_SNAKE_CASE_ = ''' ''' + output_txt SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) return output_txt, output_ids def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = tokenizer(['''hi</s>''', '''I went to the gym</s>''', '''</s>'''] ) SCREAMING_SNAKE_CASE_ = tokenizer(['''hi''', '''I went to the gym''', ''''''] ) self.assertListEqual(batch_with_eos_added['''input_ids'''] , batch_without_eos_added['''input_ids'''] ) def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = '''Unicode €.''' SCREAMING_SNAKE_CASE_ = tokenizer(_A ) SCREAMING_SNAKE_CASE_ = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A ) self.assertEqual(_A , '''Unicode €.</s>''' ) SCREAMING_SNAKE_CASE_ = tokenizer('''e è é ê ë''' ) SCREAMING_SNAKE_CASE_ = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A ) self.assertEqual(_A , '''e è é ê ë</s>''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''e è é ê ë</s>''' ) def _UpperCamelCase ( self ) -> List[str]: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off SCREAMING_SNAKE_CASE_ = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on SCREAMING_SNAKE_CASE_ = tokenizer(_A , padding=_A , return_tensors=_A ) self.assertIsInstance(_A , _A ) if FRAMEWORK != "jax": SCREAMING_SNAKE_CASE_ = list(batch.input_ids.numpy()[0] ) else: SCREAMING_SNAKE_CASE_ = list(batch.input_ids.tolist()[0] ) self.assertListEqual(_A , _A ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] SCREAMING_SNAKE_CASE_ = tokenizer(_A , padding=_A , return_tensors=_A ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , _A ) self.assertIn('''attention_mask''' , _A ) self.assertNotIn('''decoder_input_ids''' , _A ) self.assertNotIn('''decoder_attention_mask''' , _A ) def _UpperCamelCase ( self ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = [ '''Summary of the text.''', '''Another summary.''', ] SCREAMING_SNAKE_CASE_ = tokenizer( text_target=_A , max_length=32 , padding='''max_length''' , truncation=_A , return_tensors=_A ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def _UpperCamelCase ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization. </s>'''] SCREAMING_SNAKE_CASE_ = ['''Summary of the text. </s>'''] # fmt: off SCREAMING_SNAKE_CASE_ = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] SCREAMING_SNAKE_CASE_ = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on SCREAMING_SNAKE_CASE_ = tokenizer(_A , text_target=_A ) self.assertEqual(_A , batch['''input_ids'''][0] ) self.assertEqual(_A , batch['''labels'''][0] ) def _UpperCamelCase ( self ) -> Dict: # safety check on max_len default value so we are sure the test works SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = ''' He is very happy, UNwant\u00E9d,running''' SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) shutil.rmtree(_A ) SCREAMING_SNAKE_CASE_ = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) SCREAMING_SNAKE_CASE_ = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(_A ) def _UpperCamelCase ( self ) -> int: SCREAMING_SNAKE_CASE_ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) with open(os.path.join(_A , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: SCREAMING_SNAKE_CASE_ = json.load(_A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: SCREAMING_SNAKE_CASE_ = json.load(_A ) SCREAMING_SNAKE_CASE_ = [F'''<extra_id_{i}>''' for i in range(125 )] SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(_A , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained( _A , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=_A )] SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained( _A , additional_special_tokens=_A , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained(_A ) self.assertTrue(tokenizer.decode([255] ) == '''''' ) def _UpperCamelCase ( self ) -> int: pass def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> Optional[int]: pass def _UpperCamelCase ( self ) -> Union[str, Any]: # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens SCREAMING_SNAKE_CASE_ = self.get_tokenizers(fast=_A , do_lower_case=_A ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ = ['''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''x''', '''t''', '''</s>'''] SCREAMING_SNAKE_CASE_ = tokenizer.convert_tokens_to_string(_A ) self.assertIsInstance(_A , _A ) def _UpperCamelCase ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ = [ '''bos_token''', '''eos_token''', '''unk_token''', '''sep_token''', '''pad_token''', '''cls_token''', '''mask_token''', ] SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = tokenizer.convert_ids_to_tokens( _A , skip_special_tokens=_A ) for attr in attributes_list: setattr(_A , attr + '''_id''' , _A ) self.assertEqual(getattr(_A , _A ) , _A ) self.assertEqual(getattr(_A , attr + '''_id''' ) , _A ) setattr(_A , attr + '''_id''' , _A ) self.assertEqual(getattr(_A , _A ) , _A ) self.assertEqual(getattr(_A , attr + '''_id''' ) , _A ) setattr(_A , '''additional_special_tokens_ids''' , [] ) self.assertListEqual(getattr(_A , '''additional_special_tokens''' ) , [] ) self.assertListEqual(getattr(_A , '''additional_special_tokens_ids''' ) , [] ) setattr(_A , '''additional_special_tokens_ids''' , [token_id_to_test_setters] ) self.assertListEqual(getattr(_A , '''additional_special_tokens''' ) , [token_to_test_setters] ) self.assertListEqual(getattr(_A , '''additional_special_tokens_ids''' ) , [token_id_to_test_setters] )
299
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCAmelCase ={"configuration_yolos": ["YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP", "YolosConfig", "YolosOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =["YolosFeatureExtractor"] __UpperCAmelCase =["YolosImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =[ "YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST", "YolosForObjectDetection", "YolosModel", "YolosPreTrainedModel", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys __UpperCAmelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
351
'''simple docstring''' def __lowerCAmelCase ( UpperCamelCase__ ) -> str: return "".join(chr(ord(UpperCamelCase__ ) - 32 ) if '''a''' <= char <= '''z''' else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
237
0
import argparse import os import re _SCREAMING_SNAKE_CASE = "src/transformers" # Pattern that looks at the indentation in a line. _SCREAMING_SNAKE_CASE = re.compile(r"""^(\s*)\S""") # Pattern that matches `"key":" and puts `key` in group 0. _SCREAMING_SNAKE_CASE = re.compile(r"""^\s*\"([^\"]+)\":""") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _SCREAMING_SNAKE_CASE = re.compile(r"""^\s*_import_structure\[\"([^\"]+)\"\]""") # Pattern that matches `"key",` and puts `key` in group 0. _SCREAMING_SNAKE_CASE = re.compile(r"""^\s*\"([^\"]+)\",\s*$""") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _SCREAMING_SNAKE_CASE = re.compile(r"""\[([^\]]+)\]""") def lowercase( UpperCamelCase_ ) -> Dict: '''simple docstring''' UpperCamelCase = _re_indent.search(__lowerCamelCase ) return "" if search is None else search.groups()[0] def lowercase( UpperCamelCase_ , UpperCamelCase_="" , UpperCamelCase_=None , UpperCamelCase_=None ) -> List[str]: '''simple docstring''' UpperCamelCase = 0 UpperCamelCase = code.split("""\n""" ) if start_prompt is not None: while not lines[index].startswith(__lowerCamelCase ): index += 1 UpperCamelCase = ['''\n'''.join(lines[:index] )] else: UpperCamelCase = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). UpperCamelCase = [lines[index]] index += 1 while index < len(__lowerCamelCase ) and (end_prompt is None or not lines[index].startswith(__lowerCamelCase )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(__lowerCamelCase ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + """ """ ): current_block.append(lines[index] ) blocks.append("""\n""".join(__lowerCamelCase ) ) if index < len(__lowerCamelCase ) - 1: UpperCamelCase = [lines[index + 1]] index += 1 else: UpperCamelCase = [] else: blocks.append("""\n""".join(__lowerCamelCase ) ) UpperCamelCase = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(__lowerCamelCase ) > 0: blocks.append("""\n""".join(__lowerCamelCase ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(__lowerCamelCase ): blocks.append("""\n""".join(lines[index:] ) ) return blocks def lowercase( UpperCamelCase_ ) -> str: '''simple docstring''' def _inner(UpperCamelCase_ ): return key(__lowerCamelCase ).lower().replace("""_""" , """""" ) return _inner def lowercase( UpperCamelCase_ , UpperCamelCase_=None ) -> int: '''simple docstring''' # If no key is provided, we use a noop. def noop(UpperCamelCase_ ): return x if key is None: UpperCamelCase = noop # Constants are all uppercase, they go first. UpperCamelCase = [obj for obj in objects if key(__lowerCamelCase ).isupper()] # Classes are not all uppercase but start with a capital, they go second. UpperCamelCase = [obj for obj in objects if key(__lowerCamelCase )[0].isupper() and not key(__lowerCamelCase ).isupper()] # Functions begin with a lowercase, they go last. UpperCamelCase = [obj for obj in objects if not key(__lowerCamelCase )[0].isupper()] UpperCamelCase = ignore_underscore(__lowerCamelCase ) return sorted(__lowerCamelCase , key=__lowerCamelCase ) + sorted(__lowerCamelCase , key=__lowerCamelCase ) + sorted(__lowerCamelCase , key=__lowerCamelCase ) def lowercase( UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' # This inner function sort imports between [ ]. def _replace(UpperCamelCase_ ): UpperCamelCase = match.groups()[0] if "," not in imports: return f"""[{imports}]""" UpperCamelCase = [part.strip().replace("""\"""" , """""" ) for part in imports.split(""",""" )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCamelCase = keys[:-1] return "[" + ", ".join([f"""\"{k}\"""" for k in sort_objects(__lowerCamelCase )] ) + "]" UpperCamelCase = import_statement.split("""\n""" ) if len(__lowerCamelCase ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. UpperCamelCase = 2 if lines[1].strip() == '''[''' else 1 UpperCamelCase = [(i, _re_strip_line.search(__lowerCamelCase ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] UpperCamelCase = sort_objects(__lowerCamelCase , key=lambda UpperCamelCase_ : x[1] ) UpperCamelCase = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(__lowerCamelCase ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: UpperCamelCase = _re_bracket_content.sub(_replace , lines[1] ) else: UpperCamelCase = [part.strip().replace("""\"""" , """""" ) for part in lines[1].split(""",""" )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: UpperCamelCase = keys[:-1] UpperCamelCase = get_indent(lines[1] ) + ''', '''.join([f"""\"{k}\"""" for k in sort_objects(__lowerCamelCase )] ) return "\n".join(__lowerCamelCase ) else: # Finally we have to deal with imports fitting on one line UpperCamelCase = _re_bracket_content.sub(_replace , __lowerCamelCase ) return import_statement def lowercase( UpperCamelCase_ , UpperCamelCase_=True ) -> str: '''simple docstring''' with open(__lowerCamelCase , encoding="""utf-8""" ) as f: UpperCamelCase = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 UpperCamelCase = split_code_in_indented_blocks( __lowerCamelCase , start_prompt="""_import_structure = {""" , end_prompt="""if TYPE_CHECKING:""" ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(__lowerCamelCase ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. UpperCamelCase = main_blocks[block_idx] UpperCamelCase = block.split("""\n""" ) # Get to the start of the imports. UpperCamelCase = 0 while line_idx < len(__lowerCamelCase ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: UpperCamelCase = len(__lowerCamelCase ) else: line_idx += 1 if line_idx >= len(__lowerCamelCase ): continue # Ignore beginning and last line: they don't contain anything. UpperCamelCase = '''\n'''.join(block_lines[line_idx:-1] ) UpperCamelCase = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. UpperCamelCase = split_code_in_indented_blocks(__lowerCamelCase , indent_level=__lowerCamelCase ) # We have two categories of import key: list or _import_structure[key].append/extend UpperCamelCase = _re_direct_key if '''_import_structure = {''' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. UpperCamelCase = [(pattern.search(__lowerCamelCase ).groups()[0] if pattern.search(__lowerCamelCase ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. UpperCamelCase = [(i, key) for i, key in enumerate(__lowerCamelCase ) if key is not None] UpperCamelCase = [x[0] for x in sorted(__lowerCamelCase , key=lambda UpperCamelCase_ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. UpperCamelCase = 0 UpperCamelCase = [] for i in range(len(__lowerCamelCase ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: UpperCamelCase = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(__lowerCamelCase ) count += 1 # And we put our main block back together with its first and last line. UpperCamelCase = '''\n'''.join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(__lowerCamelCase ): if check_only: return True else: print(f"""Overwriting {file}.""" ) with open(__lowerCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write("""\n""".join(__lowerCamelCase ) ) def lowercase( UpperCamelCase_=True ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase = [] for root, _, files in os.walk(__lowerCamelCase ): if "__init__.py" in files: UpperCamelCase = sort_imports(os.path.join(__lowerCamelCase , """__init__.py""" ) , check_only=__lowerCamelCase ) if result: UpperCamelCase = [os.path.join(__lowerCamelCase , """__init__.py""" )] if len(__lowerCamelCase ) > 0: raise ValueError(f"""Would overwrite {len(__lowerCamelCase )} files, run `make style`.""" ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") _SCREAMING_SNAKE_CASE = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
343
"""simple docstring""" import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import VideoMAEConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEModel, ) from transformers.models.videomae.modeling_videomae import VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[Any], lowerCamelCase : Any, lowerCamelCase : List[Any]=13, lowerCamelCase : Any=10, lowerCamelCase : Optional[Any]=3, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : Dict=2, lowerCamelCase : Tuple=2, lowerCamelCase : List[str]=True, lowerCamelCase : Optional[int]=True, lowerCamelCase : Dict=32, lowerCamelCase : Any=5, lowerCamelCase : Dict=4, lowerCamelCase : Any=37, lowerCamelCase : Union[str, Any]="gelu", lowerCamelCase : Dict=0.1, lowerCamelCase : Union[str, Any]=0.1, lowerCamelCase : Dict=10, lowerCamelCase : str=0.02, lowerCamelCase : List[Any]=0.9, lowerCamelCase : List[Any]=None, )-> str: lowerCamelCase__ : List[str] =parent lowerCamelCase__ : Any =batch_size lowerCamelCase__ : str =image_size lowerCamelCase__ : Optional[Any] =num_channels lowerCamelCase__ : Optional[int] =patch_size lowerCamelCase__ : List[str] =tubelet_size lowerCamelCase__ : Optional[Any] =num_frames lowerCamelCase__ : Any =is_training lowerCamelCase__ : List[Any] =use_labels lowerCamelCase__ : Union[str, Any] =hidden_size lowerCamelCase__ : List[str] =num_hidden_layers lowerCamelCase__ : str =num_attention_heads lowerCamelCase__ : List[Any] =intermediate_size lowerCamelCase__ : Any =hidden_act lowerCamelCase__ : int =hidden_dropout_prob lowerCamelCase__ : Optional[int] =attention_probs_dropout_prob lowerCamelCase__ : Optional[Any] =type_sequence_label_size lowerCamelCase__ : int =initializer_range lowerCamelCase__ : Optional[Any] =mask_ratio lowerCamelCase__ : Any =scope # in VideoMAE, the number of tokens equals num_frames/tubelet_size * num_patches per frame lowerCamelCase__ : Optional[Any] =(image_size // patch_size) ** 2 lowerCamelCase__ : Any =(num_frames // tubelet_size) * self.num_patches_per_frame # use this variable to define bool_masked_pos lowerCamelCase__ : List[Any] =int(mask_ratio * self.seq_length ) def snake_case ( self : Dict )-> Union[str, Any]: lowerCamelCase__ : str =floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Any =None if self.use_labels: lowerCamelCase__ : Union[str, Any] =ids_tensor([self.batch_size], self.type_sequence_label_size ) lowerCamelCase__ : Optional[Any] =self.get_config() return config, pixel_values, labels def snake_case ( self : Union[str, Any] )-> Optional[int]: return VideoMAEConfig( image_size=self.image_size, patch_size=self.patch_size, num_channels=self.num_channels, num_frames=self.num_frames, tubelet_size=self.tubelet_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=lowerCamelCase, initializer_range=self.initializer_range, ) def snake_case ( self : Dict, lowerCamelCase : Tuple, lowerCamelCase : Optional[Any], lowerCamelCase : Any )-> Union[str, Any]: lowerCamelCase__ : List[str] =VideoMAEModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() lowerCamelCase__ : int =model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : Any, lowerCamelCase : str, lowerCamelCase : Optional[int], lowerCamelCase : str )-> Dict: lowerCamelCase__ : int =VideoMAEForPreTraining(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch lowerCamelCase__ : Optional[int] =torch.ones((self.num_masks,) ) lowerCamelCase__ : List[str] =torch.cat([mask, torch.zeros(self.seq_length - mask.size(0 ) )] ) lowerCamelCase__ : int =mask.expand(self.batch_size, -1 ).bool() lowerCamelCase__ : Any =model(lowerCamelCase, lowerCamelCase ) # model only returns predictions for masked patches lowerCamelCase__ : Optional[int] =mask.sum().item() lowerCamelCase__ : Dict =3 * self.tubelet_size * self.patch_size**2 self.parent.assertEqual(result.logits.shape, (self.batch_size, num_masked_patches, decoder_num_labels) ) def snake_case ( self : Optional[Any] )-> Tuple: lowerCamelCase__ : Tuple =self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Dict =config_and_inputs lowerCamelCase__ : List[str] ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' _a = ( (VideoMAEModel, VideoMAEForPreTraining, VideoMAEForVideoClassification) if is_torch_available() else () ) _a = ( {'feature-extraction': VideoMAEModel, 'video-classification': VideoMAEForVideoClassification} if is_torch_available() else {} ) _a = False _a = False _a = False _a = False def snake_case ( self : List[Any] )-> Tuple: lowerCamelCase__ : int =VideoMAEModelTester(self ) lowerCamelCase__ : Optional[int] =ConfigTester(self, config_class=lowerCamelCase, has_text_modality=lowerCamelCase, hidden_size=37 ) def snake_case ( self : Any, lowerCamelCase : List[Any], lowerCamelCase : Optional[Any], lowerCamelCase : List[str]=False )-> Tuple: lowerCamelCase__ : str =copy.deepcopy(lowerCamelCase ) if model_class == VideoMAEForPreTraining: # important: each video needs to have the same number of masked patches # hence we define a single mask, which we then repeat for each example in the batch lowerCamelCase__ : Any =torch.ones((self.model_tester.num_masks,) ) lowerCamelCase__ : Dict =torch.cat([mask, torch.zeros(self.model_tester.seq_length - mask.size(0 ) )] ) lowerCamelCase__ : Optional[int] =mask.expand(self.model_tester.batch_size, -1 ).bool() lowerCamelCase__ : int =bool_masked_pos.to(lowerCamelCase ) if return_labels: if model_class in [ *get_values(lowerCamelCase ), ]: lowerCamelCase__ : List[str] =torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCamelCase ) return inputs_dict def snake_case ( self : List[Any] )-> int: self.config_tester.run_common_tests() @unittest.skip(reason='''VideoMAE does not use inputs_embeds''' ) def snake_case ( self : List[str] )-> Tuple: pass def snake_case ( self : Union[str, Any] )-> Union[str, Any]: lowerCamelCase__ , lowerCamelCase__ : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[str] =model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) lowerCamelCase__ : Optional[Any] =model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase, nn.Linear ) ) def snake_case ( self : Optional[int] )-> Optional[Any]: lowerCamelCase__ , lowerCamelCase__ : Any =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] =model_class(lowerCamelCase ) lowerCamelCase__ : Dict =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : Tuple =[*signature.parameters.keys()] lowerCamelCase__ : List[str] =['''pixel_values'''] self.assertListEqual(arg_names[:1], lowerCamelCase ) def snake_case ( self : Tuple )-> Optional[int]: lowerCamelCase__ : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def snake_case ( self : List[Any] )-> Union[str, Any]: lowerCamelCase__ : Optional[int] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase ) @slow def snake_case ( self : List[Any] )-> Dict: for model_name in VIDEOMAE_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : str =VideoMAEModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case ( self : List[str] )-> Optional[int]: if not self.has_attentions: pass else: lowerCamelCase__ , lowerCamelCase__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase__ : Tuple =True for model_class in self.all_model_classes: lowerCamelCase__ : Any =self.model_tester.seq_length - self.model_tester.num_masks lowerCamelCase__ : Any =( num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length ) lowerCamelCase__ : Optional[int] =True lowerCamelCase__ : Optional[int] =False lowerCamelCase__ : Optional[int] =True lowerCamelCase__ : int =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : str =outputs.attentions self.assertEqual(len(lowerCamelCase ), self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCamelCase__ : Tuple =True lowerCamelCase__ : Union[str, Any] =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : List[str] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : int =outputs.attentions self.assertEqual(len(lowerCamelCase ), self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ), [self.model_tester.num_attention_heads, seq_len, seq_len], ) lowerCamelCase__ : Union[str, Any] =len(lowerCamelCase ) # Check attention is always last and order is fine lowerCamelCase__ : List[Any] =True lowerCamelCase__ : Union[str, Any] =True lowerCamelCase__ : Dict =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Any =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) self.assertEqual(out_len + 1, len(lowerCamelCase ) ) lowerCamelCase__ : Optional[Any] =outputs.attentions self.assertEqual(len(lowerCamelCase ), self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ), [self.model_tester.num_attention_heads, seq_len, seq_len], ) def snake_case ( self : str )-> int: def check_hidden_states_output(lowerCamelCase : Optional[Any], lowerCamelCase : List[str], lowerCamelCase : Optional[Any] ): lowerCamelCase__ : List[Any] =model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): lowerCamelCase__ : Optional[Any] =model(**self._prepare_for_class(lowerCamelCase, lowerCamelCase ) ) lowerCamelCase__ : Dict =outputs.hidden_states lowerCamelCase__ : Any =self.model_tester.num_hidden_layers + 1 self.assertEqual(len(lowerCamelCase ), lowerCamelCase ) lowerCamelCase__ : Any =self.model_tester.seq_length - self.model_tester.num_masks lowerCamelCase__ : str =num_visible_patches if model_class == VideoMAEForPreTraining else self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ), [seq_length, self.model_tester.hidden_size], ) lowerCamelCase__ , lowerCamelCase__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Union[str, Any] =True check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : int =True check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def snake_case ( self : Optional[int] )-> int: pass def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : int =hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) lowerCamelCase__ : str =np.load(__lowerCamelCase ) return list(__lowerCamelCase ) @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case ( self : List[str] )-> List[Any]: # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5], image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def snake_case ( self : Optional[Any] )-> Dict: lowerCamelCase__ : str =VideoMAEForVideoClassification.from_pretrained('''MCG-NJU/videomae-base-finetuned-kinetics''' ).to( lowerCamelCase ) lowerCamelCase__ : Optional[Any] =self.default_image_processor lowerCamelCase__ : List[str] =prepare_video() lowerCamelCase__ : Union[str, Any] =image_processor(lowerCamelCase, return_tensors='''pt''' ).to(lowerCamelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : Tuple =model(**lowerCamelCase ) # verify the logits lowerCamelCase__ : Union[str, Any] =torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape, lowerCamelCase ) lowerCamelCase__ : Tuple =torch.tensor([0.3_669, -0.0_688, -0.2_421] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3], lowerCamelCase, atol=1E-4 ) ) @slow def snake_case ( self : Any )-> Tuple: lowerCamelCase__ : Tuple =VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''' ).to(lowerCamelCase ) lowerCamelCase__ : Optional[int] =self.default_image_processor lowerCamelCase__ : Dict =prepare_video() lowerCamelCase__ : Dict =image_processor(lowerCamelCase, return_tensors='''pt''' ).to(lowerCamelCase ) # add boolean mask, indicating which patches to mask lowerCamelCase__ : str =hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''', filename='''bool_masked_pos.pt''' ) lowerCamelCase__ : Dict =torch.load(lowerCamelCase ) # forward pass with torch.no_grad(): lowerCamelCase__ : Union[str, Any] =model(**lowerCamelCase ) # verify the logits lowerCamelCase__ : Dict =torch.Size([1, 1408, 1536] ) lowerCamelCase__ : Union[str, Any] =torch.tensor( [[0.7_994, 0.9_612, 0.8_508], [0.7_401, 0.8_958, 0.8_302], [0.5_862, 0.7_468, 0.7_325]], device=lowerCamelCase ) self.assertEqual(outputs.logits.shape, lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3], lowerCamelCase, atol=1E-4 ) ) # verify the loss (`config.norm_pix_loss` = `True`) lowerCamelCase__ : Optional[int] =torch.tensor([0.5_142], device=lowerCamelCase ) self.assertTrue(torch.allclose(outputs.loss, lowerCamelCase, atol=1E-4 ) ) # verify the loss (`config.norm_pix_loss` = `False`) lowerCamelCase__ : Union[str, Any] =VideoMAEForPreTraining.from_pretrained('''MCG-NJU/videomae-base-short''', norm_pix_loss=lowerCamelCase ).to( lowerCamelCase ) with torch.no_grad(): lowerCamelCase__ : Union[str, Any] =model(**lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =torch.tensor(torch.tensor([0.6_469] ), device=lowerCamelCase ) self.assertTrue(torch.allclose(outputs.loss, lowerCamelCase, atol=1E-4 ) )
238
0
"""simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = range(2, 2_0 + 1) SCREAMING_SNAKE_CASE_ : Dict = [1_0**k for k in range(ks[-1] + 1)] SCREAMING_SNAKE_CASE_ : dict[int, dict[int, list[list[int]]]] = {} def _snake_case ( UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Any , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Dict ): A__ = sum(a_i[j] for j in range(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) ) A__ = sum(a_i[j] * base[j] for j in range(min(len(UpperCAmelCase_ ) , UpperCAmelCase_ ) ) ) A__ , A__ = 0, 0 A__ = n - i A__ = memo.get(UpperCAmelCase_ ) if sub_memo is not None: A__ = sub_memo.get(UpperCAmelCase_ ) if jumps is not None and len(UpperCAmelCase_ ) > 0: # find and make the largest jump without going over A__ = -1 for _k in range(len(UpperCAmelCase_ ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: A__ = _k break if max_jump >= 0: A__ , A__ , A__ = jumps[max_jump] # since the difference between jumps is cached, add c A__ = diff + c for j in range(min(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) ): A__ , A__ = divmod(UpperCAmelCase_ , 10 ) if new_c > 0: add(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) else: A__ = [] else: A__ = {c: []} A__ = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps A__ , A__ = next_term(UpperCAmelCase_ , k - 1 , i + dn , UpperCAmelCase_ ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead A__ , A__ = compute(UpperCAmelCase_ , UpperCAmelCase_ , i + dn , UpperCAmelCase_ ) diff += _diff dn += terms_jumped A__ = sub_memo[c] # keep jumps sorted by # of terms skipped A__ = 0 while j < len(UpperCAmelCase_ ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(UpperCAmelCase_ , (diff, dn, k) ) return (diff, dn) def _snake_case ( UpperCAmelCase_ : str , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str ): if i >= n: return 0, i if k > len(UpperCAmelCase_ ): a_i.extend([0 for _ in range(k - len(UpperCAmelCase_ ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) A__ = i A__ , A__ , A__ = 0, 0, 0 for j in range(len(UpperCAmelCase_ ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 A__ = ds_c + ds_b diff += addend A__ = 0 for j in range(UpperCAmelCase_ ): A__ = a_i[j] + addend A__ , A__ = divmod(UpperCAmelCase_ , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return diff, i - start_i def _snake_case ( UpperCAmelCase_ : Any , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] ): for j in range(UpperCAmelCase_ , len(UpperCAmelCase_ ) ): A__ = digits[j] + addend if s >= 10: A__ , A__ = divmod(UpperCAmelCase_ , 10 ) A__ = addend // 10 + quotient else: A__ = s A__ = addend // 10 if addend == 0: break while addend > 0: A__ , A__ = divmod(UpperCAmelCase_ , 10 ) digits.append(UpperCAmelCase_ ) def _snake_case ( UpperCAmelCase_ : int = 10**15 ): A__ = [1] A__ = 1 A__ = 0 while True: A__ , A__ = next_term(UpperCAmelCase_ , 20 , i + dn , UpperCAmelCase_ ) dn += terms_jumped if dn == n - i: break A__ = 0 for j in range(len(UpperCAmelCase_ ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(f"""{solution() = }""")
368
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_ : List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ : str = { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/config.json', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/config.json', 'funnel-transformer/medium-base': 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json', 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/config.json', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json', 'funnel-transformer/xlarge-base': 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json', } class a ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = "funnel" UpperCAmelCase = { "hidden_size": "d_model", "num_attention_heads": "n_head", } def __init__( self: str , UpperCamelCase: List[Any]=3_05_22 , UpperCamelCase: Any=[4, 4, 4] , UpperCamelCase: Dict=None , UpperCamelCase: Dict=2 , UpperCamelCase: Optional[int]=7_68 , UpperCamelCase: List[str]=12 , UpperCamelCase: Optional[Any]=64 , UpperCamelCase: str=30_72 , UpperCamelCase: Any="gelu_new" , UpperCamelCase: Optional[Any]=0.1 , UpperCamelCase: Any=0.1 , UpperCamelCase: List[str]=0.0 , UpperCamelCase: Tuple=0.1 , UpperCamelCase: Union[str, Any]=None , UpperCamelCase: Tuple=1e-9 , UpperCamelCase: Tuple="mean" , UpperCamelCase: str="relative_shift" , UpperCamelCase: Any=True , UpperCamelCase: List[Any]=True , UpperCamelCase: int=True , **UpperCamelCase: List[str] , ): """simple docstring""" A__ = vocab_size A__ = block_sizes A__ = [1] * len(UpperCamelCase ) if block_repeats is None else block_repeats assert len(UpperCamelCase ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." A__ = num_decoder_layers A__ = d_model A__ = n_head A__ = d_head A__ = d_inner A__ = hidden_act A__ = hidden_dropout A__ = attention_dropout A__ = activation_dropout A__ = initializer_range A__ = initializer_std A__ = layer_norm_eps assert pooling_type in [ "mean", "max", ], f"""Got {pooling_type} for `pooling_type` but only 'mean' and 'max' are supported.""" A__ = pooling_type assert attention_type in [ "relative_shift", "factorized", ], f"""Got {attention_type} for `attention_type` but only 'relative_shift' and 'factorized' are supported.""" A__ = attention_type A__ = separate_cls A__ = truncate_seq A__ = pool_q_only super().__init__(**UpperCamelCase ) @property def UpperCamelCase ( self: Optional[int] ): """simple docstring""" return sum(self.block_sizes ) @num_hidden_layers.setter def UpperCamelCase ( self: Union[str, Any] , UpperCamelCase: Tuple ): """simple docstring""" raise NotImplementedError( """This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.""" ) @property def UpperCamelCase ( self: Union[str, Any] ): """simple docstring""" return len(self.block_sizes ) @num_blocks.setter def UpperCamelCase ( self: Any , UpperCamelCase: Dict ): """simple docstring""" raise NotImplementedError("""This model does not support the setting of `num_blocks`. Please set `block_sizes`.""" )
69
0
def lowercase_ ( _A : Any , _A : Union[str, Any] ): """simple docstring""" lowerCamelCase__ : Tuple = [0 for i in range(r + 1 )] # nc0 = 1 lowerCamelCase__ : Optional[Any] = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. lowerCamelCase__ : Union[str, Any] = min(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=10, r=5))
184
import sys from collections import defaultdict class A_ : '''simple docstring''' def __init__( self ): lowercase = [] def SCREAMING_SNAKE_CASE__ ( self , snake_case ): return self.node_position[vertex] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = pos def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case ): if start > size // 2 - 1: return else: if 2 * start + 2 >= size: lowercase = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: lowercase = 2 * start + 1 else: lowercase = 2 * start + 2 if heap[smallest_child] < heap[start]: lowercase , lowercase = heap[smallest_child], positions[smallest_child] lowercase , lowercase = ( heap[start], positions[start], ) lowercase , lowercase = temp, tempa lowercase = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] , self.get_position(positions[start] ) ) self.set_position(positions[start] , snake_case ) self.top_to_bottom(snake_case , snake_case , snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case ): lowercase = position[index] while index != 0: lowercase = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: lowercase = heap[parent] lowercase = position[parent] self.set_position(position[parent] , snake_case ) else: lowercase = val lowercase = temp self.set_position(snake_case , snake_case ) break lowercase = parent else: lowercase = val lowercase = temp self.set_position(snake_case , 0 ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = len(snake_case ) // 2 - 1 for i in range(snake_case , -1 , -1 ): self.top_to_bottom(snake_case , snake_case , len(snake_case ) , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = positions[0] lowercase = sys.maxsize self.top_to_bottom(snake_case , 0 , len(snake_case ) , snake_case ) return temp def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = Heap() lowercase = [0] * len(__SCREAMING_SNAKE_CASE ) lowercase = [-1] * len(__SCREAMING_SNAKE_CASE ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph lowercase = [] # Heap of Distance of vertices from their neighboring vertex lowercase = [] for vertex in range(len(__SCREAMING_SNAKE_CASE ) ): distance_tv.append(sys.maxsize ) positions.append(__SCREAMING_SNAKE_CASE ) heap.node_position.append(__SCREAMING_SNAKE_CASE ) lowercase = [] lowercase = 1 lowercase = sys.maxsize for neighbor, distance in adjacency_list[0]: lowercase = 0 lowercase = distance heap.heapify(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for _ in range(1 , len(__SCREAMING_SNAKE_CASE ) ): lowercase = heap.delete_minimum(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) lowercase = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__SCREAMING_SNAKE_CASE )] ): lowercase = distance heap.bottom_to_top( __SCREAMING_SNAKE_CASE , heap.get_position(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > UpperCAmelCase = int(input('''Enter number of edges: ''').strip()) UpperCAmelCase = defaultdict(list) for _ in range(edges_number): UpperCAmelCase = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
195
0
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase = 3 , UpperCamelCase = 7 , UpperCamelCase = 1000000 ) -> int: lowerCamelCase__ : Dict = 0 lowerCamelCase__ : int = 1 for current_denominator in range(1 , limit + 1 ): lowerCamelCase__ : Union[str, Any] = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: lowerCamelCase__ : Any = current_numerator lowerCamelCase__ : str = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_000_000))
362
'''simple docstring''' from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline _A : Optional[Any] =logging.get_logger(__name__) @add_end_docstrings(_lowercase ) class _lowercase ( _lowercase ): def __init__( self: Union[str, Any] , **UpperCamelCase__: str ): super().__init__(**UpperCamelCase__ ) if self.framework != "pt": raise ValueError(F'''The {self.__class__} is only available in PyTorch.''' ) # No specific FOR_XXX available yet def __call__( self: Optional[Any] , UpperCamelCase__: Union[np.ndarray, bytes, str] , **UpperCamelCase__: int ): return super().__call__(UpperCamelCase__ , **UpperCamelCase__ ) def lowerCamelCase_ ( self: Optional[Any] , **UpperCamelCase__: int ): lowerCamelCase__ : Optional[Any] = {} if "candidate_labels" in kwargs: lowerCamelCase__ : str = kwargs["""candidate_labels"""] if "hypothesis_template" in kwargs: lowerCamelCase__ : int = kwargs["""hypothesis_template"""] return preprocess_params, {}, {} def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase__: int , UpperCamelCase__: List[str]=None , UpperCamelCase__: Dict="This is a sound of {}." ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): if audio.startswith("""http://""" ) or audio.startswith("""https://""" ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png lowerCamelCase__ : int = requests.get(UpperCamelCase__ ).content else: with open(UpperCamelCase__ , """rb""" ) as f: lowerCamelCase__ : Dict = f.read() if isinstance(UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase__ : str = ffmpeg_read(UpperCamelCase__ , self.feature_extractor.sampling_rate ) if not isinstance(UpperCamelCase__ , np.ndarray ): raise ValueError("""We expect a numpy ndarray as input""" ) if len(audio.shape ) != 1: raise ValueError("""We expect a single channel audio input for ZeroShotAudioClassificationPipeline""" ) lowerCamelCase__ : Optional[Any] = self.feature_extractor( [audio] , sampling_rate=self.feature_extractor.sampling_rate , return_tensors="""pt""" ) lowerCamelCase__ : Any = candidate_labels lowerCamelCase__ : Any = [hypothesis_template.format(UpperCamelCase__ ) for x in candidate_labels] lowerCamelCase__ : int = self.tokenizer(UpperCamelCase__ , return_tensors=self.framework , padding=UpperCamelCase__ ) lowerCamelCase__ : Tuple = [text_inputs] return inputs def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : Any = model_inputs.pop("""candidate_labels""" ) lowerCamelCase__ : Union[str, Any] = model_inputs.pop("""text_inputs""" ) if isinstance(text_inputs[0] , UpperCamelCase__ ): lowerCamelCase__ : Optional[int] = text_inputs[0] else: # Batching case. lowerCamelCase__ : Tuple = text_inputs[0][0] lowerCamelCase__ : Tuple = self.model(**UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : str = { """candidate_labels""": candidate_labels, """logits""": outputs.logits_per_audio, } return model_outputs def lowerCamelCase_ ( self: List[str] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = model_outputs.pop("""candidate_labels""" ) lowerCamelCase__ : int = model_outputs["""logits"""][0] if self.framework == "pt": lowerCamelCase__ : Optional[int] = logits.softmax(dim=0 ) lowerCamelCase__ : Any = probs.tolist() else: raise ValueError("""`tf` framework not supported.""" ) lowerCamelCase__ : Tuple = [ {"""score""": score, """label""": candidate_label} for score, candidate_label in sorted(zip(UpperCamelCase__ , UpperCamelCase__ ) , key=lambda UpperCamelCase__ : -x[0] ) ] return result
129
0
from __future__ import annotations from statistics import mean def lowerCamelCase__ ( __lowerCamelCase : list[int] , __lowerCamelCase : list[int] , __lowerCamelCase : int ): __UpperCAmelCase : Optional[Any] = [0] * no_of_processes __UpperCAmelCase : int = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(__lowerCamelCase ): __UpperCAmelCase : Optional[int] = burst_time[i] __UpperCAmelCase : list[int] = [] __UpperCAmelCase : Dict = 0 __UpperCAmelCase : Optional[Any] = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: __UpperCAmelCase : Optional[int] = [] __UpperCAmelCase : Tuple = -1 for i in range(__lowerCamelCase ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(__lowerCamelCase ) if len(__lowerCamelCase ) > 0: __UpperCAmelCase : List[str] = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: __UpperCAmelCase : Union[str, Any] = i total_time += burst_time[target_process] completed += 1 __UpperCAmelCase : List[Any] = 0 __UpperCAmelCase : Tuple = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def lowerCamelCase__ ( __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : list[int] ): __UpperCAmelCase : Optional[int] = [0] * no_of_processes for i in range(__lowerCamelCase ): __UpperCAmelCase : Union[str, Any] = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("[TEST CASE 01]") a : Dict = 4 a : Tuple = [2, 5, 3, 7] a : str = [0, 0, 0, 0] a : str = calculate_waitingtime(arrival_time, burst_time, no_of_processes) a : int = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time") for i, process_id in enumerate(list(range(1, 5))): print( f"""{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t""" f"""{waiting_time[i]}\t\t\t\t{turn_around_time[i]}""" ) print(f"""\nAverage waiting time = {mean(waiting_time):.5f}""") print(f"""Average turnaround time = {mean(turn_around_time):.5f}""")
114
import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def lowerCamelCase__ ( __lowerCamelCase : Dict , __lowerCamelCase : int ): __UpperCAmelCase : List[Any] = checkpoint __UpperCAmelCase : Tuple = {} __UpperCAmelCase : Dict = vae_state_dict["""encoder.conv_in.weight"""] __UpperCAmelCase : Optional[Any] = vae_state_dict["""encoder.conv_in.bias"""] __UpperCAmelCase : Dict = vae_state_dict["""encoder.conv_out.weight"""] __UpperCAmelCase : Union[str, Any] = vae_state_dict["""encoder.conv_out.bias"""] __UpperCAmelCase : List[Any] = vae_state_dict["""encoder.norm_out.weight"""] __UpperCAmelCase : Tuple = vae_state_dict["""encoder.norm_out.bias"""] __UpperCAmelCase : Dict = vae_state_dict["""decoder.conv_in.weight"""] __UpperCAmelCase : Tuple = vae_state_dict["""decoder.conv_in.bias"""] __UpperCAmelCase : Optional[int] = vae_state_dict["""decoder.conv_out.weight"""] __UpperCAmelCase : Optional[int] = vae_state_dict["""decoder.conv_out.bias"""] __UpperCAmelCase : Optional[Any] = vae_state_dict["""decoder.norm_out.weight"""] __UpperCAmelCase : Union[str, Any] = vae_state_dict["""decoder.norm_out.bias"""] __UpperCAmelCase : Optional[int] = vae_state_dict["""quant_conv.weight"""] __UpperCAmelCase : int = vae_state_dict["""quant_conv.bias"""] __UpperCAmelCase : Union[str, Any] = vae_state_dict["""post_quant_conv.weight"""] __UpperCAmelCase : Any = vae_state_dict["""post_quant_conv.bias"""] # Retrieves the keys for the encoder down blocks only __UpperCAmelCase : int = len({""".""".join(layer.split(""".""" )[:3] ) for layer in vae_state_dict if """encoder.down""" in layer} ) __UpperCAmelCase : Optional[Any] = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } # Retrieves the keys for the decoder up blocks only __UpperCAmelCase : Dict = len({""".""".join(layer.split(""".""" )[:3] ) for layer in vae_state_dict if """decoder.up""" in layer} ) __UpperCAmelCase : Optional[int] = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } for i in range(__lowerCamelCase ): __UpperCAmelCase : List[Any] = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: __UpperCAmelCase : Optional[Any] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) __UpperCAmelCase : int = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) __UpperCAmelCase : Optional[int] = renew_vae_resnet_paths(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = {"""old""": f"""down.{i}.block""", """new""": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) __UpperCAmelCase : Tuple = [key for key in vae_state_dict if """encoder.mid.block""" in key] __UpperCAmelCase : Optional[Any] = 2 for i in range(1 , num_mid_res_blocks + 1 ): __UpperCAmelCase : Dict = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] __UpperCAmelCase : Tuple = renew_vae_resnet_paths(__lowerCamelCase ) __UpperCAmelCase : Tuple = {"""old""": f"""mid.block_{i}""", """new""": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) __UpperCAmelCase : List[Any] = [key for key in vae_state_dict if """encoder.mid.attn""" in key] __UpperCAmelCase : str = renew_vae_attention_paths(__lowerCamelCase ) __UpperCAmelCase : List[str] = {"""old""": """mid.attn_1""", """new""": """mid_block.attentions.0"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) for i in range(__lowerCamelCase ): __UpperCAmelCase : Optional[Any] = num_up_blocks - 1 - i __UpperCAmelCase : Union[str, Any] = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: __UpperCAmelCase : int = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] __UpperCAmelCase : Dict = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] __UpperCAmelCase : Dict = renew_vae_resnet_paths(__lowerCamelCase ) __UpperCAmelCase : Optional[Any] = {"""old""": f"""up.{block_id}.block""", """new""": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) __UpperCAmelCase : Tuple = [key for key in vae_state_dict if """decoder.mid.block""" in key] __UpperCAmelCase : Union[str, Any] = 2 for i in range(1 , num_mid_res_blocks + 1 ): __UpperCAmelCase : Dict = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] __UpperCAmelCase : List[Any] = renew_vae_resnet_paths(__lowerCamelCase ) __UpperCAmelCase : int = {"""old""": f"""mid.block_{i}""", """new""": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) __UpperCAmelCase : Dict = [key for key in vae_state_dict if """decoder.mid.attn""" in key] __UpperCAmelCase : List[Any] = renew_vae_attention_paths(__lowerCamelCase ) __UpperCAmelCase : List[str] = {"""old""": """mid.attn_1""", """new""": """mid_block.attentions.0"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) return new_checkpoint def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : str , ): # Only support V1 __UpperCAmelCase : Optional[int] = requests.get( """ https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml""" ) __UpperCAmelCase : Optional[int] = io.BytesIO(r.content ) __UpperCAmelCase : Dict = OmegaConf.load(__lowerCamelCase ) __UpperCAmelCase : str = 512 __UpperCAmelCase : Any = """cuda""" if torch.cuda.is_available() else """cpu""" if checkpoint_path.endswith("""safetensors""" ): from safetensors import safe_open __UpperCAmelCase : List[Any] = {} with safe_open(__lowerCamelCase , framework="""pt""" , device="""cpu""" ) as f: for key in f.keys(): __UpperCAmelCase : str = f.get_tensor(__lowerCamelCase ) else: __UpperCAmelCase : Optional[int] = torch.load(__lowerCamelCase , map_location=__lowerCamelCase )["""state_dict"""] # Convert the VAE model. __UpperCAmelCase : Optional[int] = create_vae_diffusers_config(__lowerCamelCase , image_size=__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = custom_convert_ldm_vae_checkpoint(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Optional[int] = AutoencoderKL(**__lowerCamelCase ) vae.load_state_dict(__lowerCamelCase ) vae.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": a : List[Any] = argparse.ArgumentParser() parser.add_argument("--vae_pt_path", default=None, type=str, required=True, help="Path to the VAE.pt to convert.") parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the VAE.pt to convert.") a : Optional[int] = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
114
1
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig lowerCAmelCase_ = logging.get_logger(__name__) class _A : def __init__( self : List[str] , _A : Any , _A : str ) -> Tuple: """simple docstring""" lowercase : List[Any] = question_encoder lowercase : Any = generator lowercase : Optional[int] = self.question_encoder def __a ( self : str , _A : str ) -> Union[str, Any]: """simple docstring""" if os.path.isfile(_lowerCAmelCase ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) lowercase : Union[str, Any] = os.path.join(_lowerCAmelCase , '''question_encoder_tokenizer''' ) lowercase : Tuple = os.path.join(_lowerCAmelCase , '''generator_tokenizer''' ) self.question_encoder.save_pretrained(_lowerCAmelCase ) self.generator.save_pretrained(_lowerCAmelCase ) @classmethod def __a ( cls : Union[str, Any] , _A : Tuple , **_A : List[str] ) -> Tuple: """simple docstring""" from ..auto.tokenization_auto import AutoTokenizer lowercase : List[Any] = kwargs.pop('''config''' , _lowerCAmelCase ) if config is None: lowercase : Any = RagConfig.from_pretrained(_lowerCAmelCase ) lowercase : Optional[Any] = AutoTokenizer.from_pretrained( _lowerCAmelCase , config=config.question_encoder , subfolder='''question_encoder_tokenizer''' ) lowercase : Optional[int] = AutoTokenizer.from_pretrained( _lowerCAmelCase , config=config.generator , subfolder='''generator_tokenizer''' ) return cls(question_encoder=_lowerCAmelCase , generator=_lowerCAmelCase ) def __call__( self : List[str] , *_A : str , **_A : Union[str, Any] ) -> Any: """simple docstring""" return self.current_tokenizer(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self : Union[str, Any] , *_A : Optional[int] , **_A : Optional[int] ) -> Union[str, Any]: """simple docstring""" return self.generator.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self : int , *_A : Tuple , **_A : List[str] ) -> Optional[int]: """simple docstring""" return self.generator.decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self : Optional[int] ) -> Optional[int]: """simple docstring""" lowercase : Any = self.question_encoder def __a ( self : Optional[Any] ) -> str: """simple docstring""" lowercase : Dict = self.generator def __a ( self : Any , _A : List[str] , _A : Optional[List[str]] = None , _A : Optional[int] = None , _A : Optional[int] = None , _A : str = "longest" , _A : str = None , _A : bool = True , **_A : List[str] , ) -> str: """simple docstring""" warnings.warn( '''`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ''' '''regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ''' '''context manager to prepare your targets. See the documentation of your specific tokenizer for more ''' '''details''' , _lowerCAmelCase , ) if max_length is None: lowercase : str = self.current_tokenizer.model_max_length lowercase : Any = self( _lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_tensors=_lowerCAmelCase , max_length=_lowerCAmelCase , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , **_lowerCAmelCase , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: lowercase : str = self.current_tokenizer.model_max_length lowercase : Optional[Any] = self( text_target=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , return_tensors=_lowerCAmelCase , padding=_lowerCAmelCase , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , **_lowerCAmelCase , ) lowercase : Optional[int] = labels["""input_ids"""] return model_inputs
370
from typing import Union import fire import torch from tqdm import tqdm def snake_case( __magic_name__ , __magic_name__ = "cpu" , __magic_name__ = None ) -> None: '''simple docstring''' lowercase : int = torch.load(__magic_name__ , map_location=__magic_name__ ) for k, v in tqdm(state_dict.items() ): if not isinstance(__magic_name__ , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) lowercase : Any = v.half() if save_path is None: # overwrite src_path lowercase : str = src_path torch.save(__magic_name__ , __magic_name__ ) if __name__ == "__main__": fire.Fire(convert)
116
0
import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : List[str] , lowerCAmelCase_ : List[Any] ) -> Optional[Any]: __lowerCAmelCase = 3 __lowerCAmelCase = 2_5_0 __lowerCAmelCase = ids_tensor((batch_size, length) , lowerCAmelCase_ ) __lowerCAmelCase = torch.ones((batch_size, length) , device=lowerCAmelCase_ , dtype=torch.float ) / length return input_ids, scores def lowercase ( self : List[Any] ) -> Any: __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) __lowerCAmelCase = StoppingCriteriaList( [ MaxLengthCriteria(max_length=1_0 ), MaxTimeCriteria(max_time=0.1 ), ] ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(1_0 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[str] ) -> Union[str, Any]: __lowerCAmelCase = MaxLengthCriteria(max_length=1_0 ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(1_0 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5 ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(9 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(1_0 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase = StoppingCriteriaList([criteria] ) self.assertEqual(criteria_list.max_length , 1_0 ) def lowercase ( self : Dict ) -> int: __lowerCAmelCase , __lowerCAmelCase = self._get_tensors(5 ) __lowerCAmelCase = MaxTimeCriteria(max_time=0.1 ) self.assertFalse(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) __lowerCAmelCase = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2 ) self.assertTrue(criteria(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : List[str] ) -> List[Any]: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_0 ) with self.assertWarns(lowerCAmelCase_ ): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(1_0 )] ) , 1_1 ) __lowerCAmelCase = validate_stopping_criteria(StoppingCriteriaList() , 1_1 ) self.assertEqual(len(lowerCAmelCase_ ) , 1 )
284
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _snake_case : Tuple = '\\n@article{hendrycksmath2021,\n title={Measuring Mathematical Problem Solving With the MATH Dataset},\n author={Dan Hendrycks\n and Collin Burns\n and Saurav Kadavath\n and Akul Arora\n and Steven Basart\n and Eric Tang\n and Dawn Song\n and Jacob Steinhardt},\n journal={arXiv preprint arXiv:2103.03874},\n year={2021}\n}\n' _snake_case : str = '\\nThis metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset.\nIt first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy.\n' _snake_case : List[str] = R'\nCalculates accuracy after canonicalizing inputs.\n\nArgs:\n predictions: list of predictions to score. Each prediction\n is a string that contains natural language and LaTex.\n references: list of reference for each prediction. Each\n reference is a string that contains natural language\n and LaTex.\nReturns:\n accuracy: accuracy after canonicalizing inputs\n (e.g., converting "1/2" to "\\frac{1}{2}")\n\nExamples:\n >>> metric = datasets.load_metric("competition_math")\n >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"])\n >>> print(results)\n {\'accuracy\': 1.0}\n' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase ( self : str ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def lowercase ( self : str , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[int] ) -> List[Any]: __lowerCAmelCase = 0.0 for i, j in zip(lowerCAmelCase_ , lowerCAmelCase_ ): n_correct += 1.0 if math_equivalence.is_equiv(lowerCAmelCase_ , lowerCAmelCase_ ) else 0.0 __lowerCAmelCase = n_correct / len(lowerCAmelCase_ ) return { "accuracy": accuracy, }
284
1
import argparse import struct import unittest class __snake_case : def __init__( self , snake_case__ ) -> None: '''simple docstring''' UpperCAmelCase : Tuple =data # Initialize hash values UpperCAmelCase : int =[ 0x6_A_0_9_E_6_6_7, 0xB_B_6_7_A_E_8_5, 0x3_C_6_E_F_3_7_2, 0xA_5_4_F_F_5_3_A, 0x5_1_0_E_5_2_7_F, 0x9_B_0_5_6_8_8_C, 0x1_F_8_3_D_9_A_B, 0x5_B_E_0_C_D_1_9, ] # Initialize round constants UpperCAmelCase : Tuple =[ 0x4_2_8_A_2_F_9_8, 0x7_1_3_7_4_4_9_1, 0xB_5_C_0_F_B_C_F, 0xE_9_B_5_D_B_A_5, 0x3_9_5_6_C_2_5_B, 0x5_9_F_1_1_1_F_1, 0x9_2_3_F_8_2_A_4, 0xA_B_1_C_5_E_D_5, 0xD_8_0_7_A_A_9_8, 0x1_2_8_3_5_B_0_1, 0x2_4_3_1_8_5_B_E, 0x5_5_0_C_7_D_C_3, 0x7_2_B_E_5_D_7_4, 0x8_0_D_E_B_1_F_E, 0x9_B_D_C_0_6_A_7, 0xC_1_9_B_F_1_7_4, 0xE_4_9_B_6_9_C_1, 0xE_F_B_E_4_7_8_6, 0x0_F_C_1_9_D_C_6, 0x2_4_0_C_A_1_C_C, 0x2_D_E_9_2_C_6_F, 0x4_A_7_4_8_4_A_A, 0x5_C_B_0_A_9_D_C, 0x7_6_F_9_8_8_D_A, 0x9_8_3_E_5_1_5_2, 0xA_8_3_1_C_6_6_D, 0xB_0_0_3_2_7_C_8, 0xB_F_5_9_7_F_C_7, 0xC_6_E_0_0_B_F_3, 0xD_5_A_7_9_1_4_7, 0x0_6_C_A_6_3_5_1, 0x1_4_2_9_2_9_6_7, 0x2_7_B_7_0_A_8_5, 0x2_E_1_B_2_1_3_8, 0x4_D_2_C_6_D_F_C, 0x5_3_3_8_0_D_1_3, 0x6_5_0_A_7_3_5_4, 0x7_6_6_A_0_A_B_B, 0x8_1_C_2_C_9_2_E, 0x9_2_7_2_2_C_8_5, 0xA_2_B_F_E_8_A_1, 0xA_8_1_A_6_6_4_B, 0xC_2_4_B_8_B_7_0, 0xC_7_6_C_5_1_A_3, 0xD_1_9_2_E_8_1_9, 0xD_6_9_9_0_6_2_4, 0xF_4_0_E_3_5_8_5, 0x1_0_6_A_A_0_7_0, 0x1_9_A_4_C_1_1_6, 0x1_E_3_7_6_C_0_8, 0x2_7_4_8_7_7_4_C, 0x3_4_B_0_B_C_B_5, 0x3_9_1_C_0_C_B_3, 0x4_E_D_8_A_A_4_A, 0x5_B_9_C_C_A_4_F, 0x6_8_2_E_6_F_F_3, 0x7_4_8_F_8_2_E_E, 0x7_8_A_5_6_3_6_F, 0x8_4_C_8_7_8_1_4, 0x8_C_C_7_0_2_0_8, 0x9_0_B_E_F_F_F_A, 0xA_4_5_0_6_C_E_B, 0xB_E_F_9_A_3_F_7, 0xC_6_7_1_7_8_F_2, ] UpperCAmelCase : int =self.preprocessing(self.data ) self.final_hash() @staticmethod def UpperCAmelCase__ ( snake_case__ ) -> bytes: '''simple docstring''' UpperCAmelCase : Any =b'''\x80''' + (b'''\x00''' * (63 - (len(snake_case__ ) + 8) % 64)) UpperCAmelCase : List[str] =struct.pack('''>Q''' , (len(snake_case__ ) * 8) ) return data + padding + big_endian_integer def UpperCAmelCase__ ( self ) -> None: '''simple docstring''' UpperCAmelCase : List[str] =[ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers UpperCAmelCase : Union[str, Any] =list(struct.unpack('''>16L''' , snake_case__ ) ) # add 48 0-ed integers words += [0] * 48 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[Any] =self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array UpperCAmelCase : Union[str, Any] =( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) UpperCAmelCase : Union[str, Any] =( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) UpperCAmelCase : List[Any] =( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_0_0_0_0_0_0_0_0 # Compression UpperCAmelCase : Optional[Any] =self.ror(snake_case__ , 6 ) ^ self.ror(snake_case__ , 11 ) ^ self.ror(snake_case__ , 25 ) UpperCAmelCase : int =(e & f) ^ ((~e & 0xF_F_F_F_F_F_F_F) & g) UpperCAmelCase : str =( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_0_0_0_0_0_0_0_0 UpperCAmelCase : List[Any] =self.ror(snake_case__ , 2 ) ^ self.ror(snake_case__ , 13 ) ^ self.ror(snake_case__ , 22 ) UpperCAmelCase : Any =(a & b) ^ (a & c) ^ (b & c) UpperCAmelCase : str =(sa + maj) % 0x1_0_0_0_0_0_0_0_0 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[Any] =( g, f, e, ((d + tempa) % 0x1_0_0_0_0_0_0_0_0), c, b, a, ((tempa + tempa) % 0x1_0_0_0_0_0_0_0_0), ) UpperCAmelCase : int =[a, b, c, d, e, f, g, h] # Modify final values UpperCAmelCase : Optional[int] =[ ((element + mutated_hash_values[index]) % 0x1_0_0_0_0_0_0_0_0) for index, element in enumerate(self.hashes ) ] UpperCAmelCase : Dict =''''''.join([hex(snake_case__ )[2:].zfill(8 ) for value in self.hashes] ) def UpperCAmelCase__ ( self , snake_case__ , snake_case__ ) -> int: '''simple docstring''' return 0xF_F_F_F_F_F_F_F & (value << (32 - rotations)) | (value >> rotations) class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> None: '''simple docstring''' import hashlib UpperCAmelCase : List[str] =bytes('''Test String''' , '''utf-8''' ) self.assertEqual(SHAaaa(snake_case__ ).hash , hashlib.shaaaa(snake_case__ ).hexdigest() ) def lowerCAmelCase_ ( )-> None: '''simple docstring''' import doctest doctest.testmod() UpperCAmelCase : Any =argparse.ArgumentParser() parser.add_argument( '''-s''' , '''--string''' , dest='''input_string''' , default='''Hello World!! Welcome to Cryptography''' , help='''Hash the string''' , ) parser.add_argument( '''-f''' , '''--file''' , dest='''input_file''' , help='''Hash contents of a file''' ) UpperCAmelCase : int =parser.parse_args() UpperCAmelCase : Tuple =args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , '''rb''' ) as f: UpperCAmelCase : Dict =f.read() else: UpperCAmelCase : Dict =bytes(__lowerCAmelCase , '''utf-8''' ) print(SHAaaa(__lowerCAmelCase ).hash ) if __name__ == "__main__": main()
78
import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( lowerCamelCase__ , unittest.TestCase ): __lowerCamelCase : List[Any] = KandinskyVaaControlnetPipeline __lowerCamelCase : int = ["""image_embeds""", """negative_image_embeds""", """hint"""] __lowerCamelCase : Optional[int] = ["""image_embeds""", """negative_image_embeds""", """hint"""] __lowerCamelCase : Optional[Any] = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __lowerCamelCase : Dict = False @property def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' return 32 @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return 32 @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.time_input_dim @property def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' return self.time_input_dim * 4 @property def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' return 100 @property def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase : Any ={ '''in_channels''': 8, # Out channels is double in channels because predicts mean and variance '''out_channels''': 8, '''addition_embed_type''': '''image_hint''', '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''encoder_hid_dim''': self.text_embedder_hidden_size, '''encoder_hid_dim_type''': '''image_proj''', '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': None, } UpperCAmelCase : List[Any] =UNetaDConditionModel(**snake_case__ ) return model @property def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) UpperCAmelCase : Any =VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : List[str] =self.dummy_unet UpperCAmelCase : Tuple =self.dummy_movq UpperCAmelCase : Union[str, Any] =DDIMScheduler( num_train_timesteps=1000 , beta_schedule='''linear''' , beta_start=0.0_0085 , beta_end=0.012 , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , steps_offset=1 , prediction_type='''epsilon''' , thresholding=snake_case__ , ) UpperCAmelCase : Tuple ={ '''unet''': unet, '''scheduler''': scheduler, '''movq''': movq, } return components def UpperCAmelCase__ ( self , snake_case__ , snake_case__=0 ) -> Any: '''simple docstring''' UpperCAmelCase : str =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) UpperCAmelCase : Tuple =floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( snake_case__ ) # create hint UpperCAmelCase : Tuple =floats_tensor((1, 3, 64, 64) , rng=random.Random(snake_case__ ) ).to(snake_case__ ) if str(snake_case__ ).startswith('''mps''' ): UpperCAmelCase : Optional[int] =torch.manual_seed(snake_case__ ) else: UpperCAmelCase : int =torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase : List[str] ={ '''image_embeds''': image_embeds, '''negative_image_embeds''': negative_image_embeds, '''hint''': hint, '''generator''': generator, '''height''': 64, '''width''': 64, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : List[Any] ='''cpu''' UpperCAmelCase : List[Any] =self.get_dummy_components() UpperCAmelCase : Tuple =self.pipeline_class(**snake_case__ ) UpperCAmelCase : Tuple =pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : Optional[int] =pipe(**self.get_dummy_inputs(snake_case__ ) ) UpperCAmelCase : str =output.images UpperCAmelCase : List[str] =pipe( **self.get_dummy_inputs(snake_case__ ) , return_dict=snake_case__ , )[0] UpperCAmelCase : Union[str, Any] =image[0, -3:, -3:, -1] UpperCAmelCase : List[str] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Union[str, Any] =np.array( [0.695_9826, 0.86_8279, 0.755_8092, 0.6876_9467, 0.8580_5804, 0.6597_7496, 0.4488_5302, 0.595_9111, 0.425_1595] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' UpperCAmelCase : Union[str, Any] =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy''' ) UpperCAmelCase : Tuple =load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/kandinskyv22/hint_image_cat.png''' ) UpperCAmelCase : int =torch.from_numpy(np.array(snake_case__ ) ).float() / 255.0 UpperCAmelCase : List[str] =hint.permute(2 , 0 , 1 ).unsqueeze(0 ) UpperCAmelCase : Dict =KandinskyVaaPriorPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-prior''' , torch_dtype=torch.floataa ) pipe_prior.to(snake_case__ ) UpperCAmelCase : int =KandinskyVaaControlnetPipeline.from_pretrained( '''kandinsky-community/kandinsky-2-2-controlnet-depth''' , torch_dtype=torch.floataa ) UpperCAmelCase : str =pipeline.to(snake_case__ ) pipeline.set_progress_bar_config(disable=snake_case__ ) UpperCAmelCase : int ='''A robot, 4k photo''' UpperCAmelCase : int =torch.Generator(device='''cuda''' ).manual_seed(0 ) UpperCAmelCase , UpperCAmelCase : List[str] =pipe_prior( snake_case__ , generator=snake_case__ , num_inference_steps=5 , negative_prompt='''''' , ).to_tuple() UpperCAmelCase : List[str] =torch.Generator(device='''cuda''' ).manual_seed(0 ) UpperCAmelCase : Dict =pipeline( image_embeds=snake_case__ , negative_image_embeds=snake_case__ , hint=snake_case__ , generator=snake_case__ , num_inference_steps=100 , output_type='''np''' , ) UpperCAmelCase : List[Any] =output.images[0] assert image.shape == (512, 512, 3) assert_mean_pixel_difference(snake_case__ , snake_case__ )
78
1
"""simple docstring""" from maths.prime_factors import prime_factors def __UpperCAmelCase ( UpperCAmelCase_ : int ) -> int: '''simple docstring''' if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): __snake_case : List[Any] = F"Input value of [number={number}] must be an integer" raise TypeError(UpperCAmelCase_ ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(UpperCAmelCase_ ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
172
"""simple docstring""" import flax.linen as nn import jax import jax.numpy as jnp class UpperCamelCase ( nn.Module ): UpperCAmelCase : int UpperCAmelCase : jnp.dtype = jnp.floataa def _lowercase (self : Any) -> Optional[int]: __snake_case : str = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__(self : Any , _A : Any) -> str: __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = hidden_states.shape __snake_case : Union[str, Any] = jax.image.resize( _A , shape=(batch, height * 2, width * 2, channels) , method='nearest' , ) __snake_case : List[Any] = self.conv(_A) return hidden_states class UpperCamelCase ( nn.Module ): UpperCAmelCase : int UpperCAmelCase : jnp.dtype = jnp.floataa def _lowercase (self : Optional[Any]) -> List[Any]: __snake_case : Dict = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__(self : int , _A : str) -> Any: # pad = ((0, 0), (0, 1), (0, 1), (0, 0)) # pad height and width dim # hidden_states = jnp.pad(hidden_states, pad_width=pad) __snake_case : Union[str, Any] = self.conv(_A) return hidden_states class UpperCamelCase ( nn.Module ): UpperCAmelCase : int UpperCAmelCase : int = None UpperCAmelCase : float = 0.0 UpperCAmelCase : bool = None UpperCAmelCase : jnp.dtype = jnp.floataa def _lowercase (self : List[str]) -> Dict: __snake_case : str = self.in_channels if self.out_channels is None else self.out_channels __snake_case : Optional[int] = nn.GroupNorm(num_groups=32 , epsilon=1E-5) __snake_case : str = nn.Conv( _A , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) __snake_case : Optional[int] = nn.Dense(_A , dtype=self.dtype) __snake_case : int = nn.GroupNorm(num_groups=32 , epsilon=1E-5) __snake_case : str = nn.Dropout(self.dropout_prob) __snake_case : Dict = nn.Conv( _A , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) __snake_case : List[str] = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut __snake_case : Optional[Any] = None if use_nin_shortcut: __snake_case : List[str] = nn.Conv( _A , kernel_size=(1, 1) , strides=(1, 1) , padding='VALID' , dtype=self.dtype , ) def __call__(self : List[Any] , _A : Union[str, Any] , _A : str , _A : int=True) -> Any: __snake_case : List[Any] = hidden_states __snake_case : Optional[Any] = self.norma(_A) __snake_case : int = nn.swish(_A) __snake_case : Optional[int] = self.conva(_A) __snake_case : Dict = self.time_emb_proj(nn.swish(_A)) __snake_case : List[str] = jnp.expand_dims(jnp.expand_dims(_A , 1) , 1) __snake_case : Any = hidden_states + temb __snake_case : Tuple = self.norma(_A) __snake_case : Dict = nn.swish(_A) __snake_case : Union[str, Any] = self.dropout(_A , _A) __snake_case : Union[str, Any] = self.conva(_A) if self.conv_shortcut is not None: __snake_case : List[Any] = self.conv_shortcut(_A) return hidden_states + residual
172
1
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union UpperCAmelCase : int = TypeVar('T') UpperCAmelCase : Optional[int] = Union[List[T], Tuple[T, ...]] UpperCAmelCase : Optional[int] = Union[T, List[T], Dict[str, T]] UpperCAmelCase : Optional[Any] = Union[str, bytes, os.PathLike]
331
'''simple docstring''' def a__ ( a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE = len(a__ ) while cur > 1: # Find the maximum number in arr __SCREAMING_SNAKE_CASE = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __SCREAMING_SNAKE_CASE = arr[mi::-1] + arr[mi + 1 : len(a__ )] # Reverse whole list __SCREAMING_SNAKE_CASE = arr[cur - 1 :: -1] + arr[cur : len(a__ )] cur -= 1 return arr if __name__ == "__main__": UpperCAmelCase : Tuple = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase : str = [int(item) for item in user_input.split(',')] print(pancake_sort(unsorted))
331
1
"""simple docstring""" def a__ ( snake_case__ = 1_00_00_00 ) -> int: lowerCamelCase = 1 lowerCamelCase = 1 lowerCamelCase = {1: 1} for inputa in range(2 , lowercase_ ): lowerCamelCase = 0 lowerCamelCase = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: lowerCamelCase = (3 * number) + 1 counter += 1 if inputa not in counters: lowerCamelCase = counter if counter > pre_counter: lowerCamelCase = inputa lowerCamelCase = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
291
def UpperCamelCase (lowercase_: int = 10 ) -> str: if not isinstance(lowercase_ , lowercase_ ) or n < 0: raise ValueError("""Invalid input""" ) A__ : List[str] = 10**n A__ : Any = 28433 * (pow(2 , 7830457 , lowercase_ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f'''{solution(10) = }''')
192
0
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class UpperCamelCase ( unittest.TestCase ): """simple docstring""" def lowerCamelCase__ ( self ,UpperCAmelCase_ ): for model_result in results.values(): for batch_size, sequence_length in zip(model_result["""bs"""] ,model_result["""ss"""] ): _lowercase : Union[str, Any] = model_result["""result"""][batch_size][sequence_length] self.assertIsNotNone(UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : Optional[int] = """sshleifer/tiny-gpt2""" _lowercase : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,eager_mode=UpperCAmelCase_ ,multi_process=UpperCAmelCase_ ,) _lowercase : Any = TensorFlowBenchmark(UpperCAmelCase_ ) _lowercase : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase__ ( self ): _lowercase : int = """sgugger/tiny-distilbert-classification""" _lowercase : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=UpperCAmelCase_ ,only_pretrain_model=UpperCAmelCase_ ,) _lowercase : List[Any] = TensorFlowBenchmark(UpperCAmelCase_ ) _lowercase : str = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase__ ( self ): _lowercase : List[str] = """sshleifer/tiny-gpt2""" _lowercase : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=UpperCAmelCase_ ,) _lowercase : List[str] = TensorFlowBenchmark(UpperCAmelCase_ ) _lowercase : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase__ ( self ): _lowercase : List[Any] = """sshleifer/tiny-gpt2""" _lowercase : str = AutoConfig.from_pretrained(UpperCAmelCase_ ) _lowercase : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,eager_mode=UpperCAmelCase_ ,multi_process=UpperCAmelCase_ ,) _lowercase : str = TensorFlowBenchmark(UpperCAmelCase_ ,[config] ) _lowercase : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase__ ( self ): _lowercase : Optional[int] = """sshleifer/tiny-gpt2""" _lowercase : str = AutoConfig.from_pretrained(UpperCAmelCase_ ) _lowercase : List[str] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=UpperCAmelCase_ ,) _lowercase : Union[str, Any] = TensorFlowBenchmark(UpperCAmelCase_ ,[config] ) _lowercase : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase__ ( self ): _lowercase : Optional[Any] = """sshleifer/tiny-gpt2""" _lowercase : Union[str, Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=UpperCAmelCase_ ,) _lowercase : Optional[int] = TensorFlowBenchmark(UpperCAmelCase_ ) _lowercase : int = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase__ ( self ): _lowercase : Tuple = """sshleifer/tiny-gpt2""" _lowercase : Tuple = AutoConfig.from_pretrained(UpperCAmelCase_ ) _lowercase : List[str] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=UpperCAmelCase_ ,) _lowercase : Dict = TensorFlowBenchmark(UpperCAmelCase_ ,[config] ) _lowercase : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def lowerCamelCase__ ( self ): _lowercase : Tuple = """patrickvonplaten/t5-tiny-random""" _lowercase : Tuple = AutoConfig.from_pretrained(UpperCAmelCase_ ) _lowercase : Dict = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=UpperCAmelCase_ ,) _lowercase : str = TensorFlowBenchmark(UpperCAmelCase_ ,configs=[config] ) _lowercase : Union[str, Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices("""GPU""" ) ) == 0 ,"""Cannot do xla on CPU.""" ) def lowerCamelCase__ ( self ): _lowercase : Any = """sshleifer/tiny-gpt2""" _lowercase : Optional[Any] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,training=UpperCAmelCase_ ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,use_xla=UpperCAmelCase_ ,multi_process=UpperCAmelCase_ ,) _lowercase : List[str] = TensorFlowBenchmark(UpperCAmelCase_ ) _lowercase : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def lowerCamelCase__ ( self ): _lowercase : Optional[int] = """sshleifer/tiny-gpt2""" with tempfile.TemporaryDirectory() as tmp_dir: _lowercase : str = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,inference=UpperCAmelCase_ ,save_to_csv=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,inference_time_csv_file=os.path.join(UpperCAmelCase_ ,"""inf_time.csv""" ) ,inference_memory_csv_file=os.path.join(UpperCAmelCase_ ,"""inf_mem.csv""" ) ,env_info_csv_file=os.path.join(UpperCAmelCase_ ,"""env.csv""" ) ,multi_process=UpperCAmelCase_ ,) _lowercase : Tuple = TensorFlowBenchmark(UpperCAmelCase_ ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCAmelCase_ ,"""inf_time.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase_ ,"""inf_mem.csv""" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCAmelCase_ ,"""env.csv""" ) ).exists() ) def lowerCamelCase__ ( self ): _lowercase : Dict = """sshleifer/tiny-gpt2""" def _check_summary_is_not_empty(UpperCAmelCase_ ): self.assertTrue(hasattr(UpperCAmelCase_ ,"""sequential""" ) ) self.assertTrue(hasattr(UpperCAmelCase_ ,"""cumulative""" ) ) self.assertTrue(hasattr(UpperCAmelCase_ ,"""current""" ) ) self.assertTrue(hasattr(UpperCAmelCase_ ,"""total""" ) ) with tempfile.TemporaryDirectory() as tmp_dir: _lowercase : Optional[int] = TensorFlowBenchmarkArguments( models=[MODEL_ID] ,inference=UpperCAmelCase_ ,sequence_lengths=[8] ,batch_sizes=[1] ,log_filename=os.path.join(UpperCAmelCase_ ,"""log.txt""" ) ,log_print=UpperCAmelCase_ ,trace_memory_line_by_line=UpperCAmelCase_ ,eager_mode=UpperCAmelCase_ ,multi_process=UpperCAmelCase_ ,) _lowercase : Optional[Any] = TensorFlowBenchmark(UpperCAmelCase_ ) _lowercase : int = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(UpperCAmelCase_ ,"""log.txt""" ) ).exists() )
336
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( ): return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] UpperCAmelCase: Any = generate_large_matrix() UpperCAmelCase: Dict = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): assert all(row == sorted(__UpperCAmelCase , reverse=__UpperCAmelCase ) for row in grid ) assert all(list(__UpperCAmelCase ) == sorted(__UpperCAmelCase , reverse=__UpperCAmelCase ) for col in zip(*__UpperCAmelCase ) ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Tuple = 0 _lowercase : List[Any] = len(__UpperCAmelCase ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: _lowercase : Tuple = (left + right) // 2 _lowercase : List[Any] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: _lowercase : Dict = mid + 1 else: _lowercase : Dict = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Any = 0 _lowercase : Optional[int] = len(grid[0] ) for i in range(len(__UpperCAmelCase ) ): _lowercase : Union[str, Any] = find_negative_index(grid[i][:bound] ) total += bound return (len(__UpperCAmelCase ) * len(grid[0] )) - total def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): return len([number for row in grid for number in row if number < 0] ) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): _lowercase : Tuple = 0 for row in grid: for i, number in enumerate(__UpperCAmelCase ): if number < 0: total += len(__UpperCAmelCase ) - i break return total def __SCREAMING_SNAKE_CASE ( ): from timeit import timeit print("""Running benchmarks""" ) _lowercase : Tuple = ( """from __main__ import count_negatives_binary_search, """ """count_negatives_brute_force, count_negatives_brute_force_with_break, grid""" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): _lowercase : Dict = timeit(F"""{func}(grid=grid)""" , setup=__UpperCAmelCase , number=500 ) print(F"""{func}() took {time:0.4f} seconds""" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
336
1
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Dict = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") UpperCamelCase : Any = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) UpperCamelCase : str = model.state_dict() def to_tf_var_name(_lowerCAmelCase ): for patt, repl in iter(_lowerCAmelCase ): UpperCamelCase : str = name.replace(_lowerCAmelCase , _lowerCAmelCase ) return F"""bert/{name}""" def create_tf_var(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCamelCase : str = tf.dtypes.as_dtype(tensor.dtype ) UpperCamelCase : str = tf.get_variable(dtype=_lowerCAmelCase , shape=tensor.shape , name=_lowerCAmelCase , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(_lowerCAmelCase ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: UpperCamelCase : str = to_tf_var_name(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): UpperCamelCase : Any = torch_tensor.T UpperCamelCase : Any = create_tf_var(tensor=_lowerCAmelCase , name=_lowerCAmelCase , session=_lowerCAmelCase ) tf.keras.backend.set_value(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : List[str] = session.run(_lowerCAmelCase ) print(F"""Successfully created {tf_name}: {np.allclose(_lowerCAmelCase , _lowerCAmelCase )}""" ) UpperCamelCase : Optional[Any] = tf.train.Saver(tf.trainable_variables() ) saver.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , model_name.replace("-" , "_" ) + ".ckpt" ) ) def A_ ( _lowerCAmelCase=None ) -> Dict: UpperCamelCase : str = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_lowerCAmelCase , required=_lowerCAmelCase , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=_lowerCAmelCase , default=_lowerCAmelCase , required=_lowerCAmelCase , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=_lowerCAmelCase , required=_lowerCAmelCase , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=_lowerCAmelCase , required=_lowerCAmelCase , help="Directory in which to save tensorflow model" ) UpperCamelCase : List[Any] = parser.parse_args(_lowerCAmelCase ) UpperCamelCase : Dict = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=_lowerCAmelCase , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
52
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self : str ) -> List[Any]: lowerCAmelCase = torch.nn.Linear(1_0 , 1_0 ) lowerCAmelCase = torch.optim.SGD(model.parameters() , 0.1 ) lowerCAmelCase = Accelerator() lowerCAmelCase = accelerator.prepare(UpperCAmelCase__ ) try: pickle.loads(pickle.dumps(UpperCAmelCase__ ) ) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''' ) AcceleratorState._reset_state()
4
0
'''simple docstring''' import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class _lowercase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): a = StableUnCLIPPipeline a = TEXT_TO_IMAGE_PARAMS a = TEXT_TO_IMAGE_BATCH_PARAMS a = TEXT_TO_IMAGE_IMAGE_PARAMS a = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false a = False def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Tuple = 32 lowerCamelCase__ : int = embedder_hidden_size # prior components torch.manual_seed(0 ) lowerCamelCase__ : Dict = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) torch.manual_seed(0 ) lowerCamelCase__ : Union[str, Any] = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=__lowercase , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) ) torch.manual_seed(0 ) lowerCamelCase__ : Any = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=__lowercase , num_layers=1 , ) torch.manual_seed(0 ) lowerCamelCase__ : Any = DDPMScheduler( variance_type="""fixed_small_log""" , prediction_type="""sample""" , num_train_timesteps=1_000 , clip_sample=__lowercase , clip_sample_range=5.0 , beta_schedule="""squaredcos_cap_v2""" , ) # regular denoising components torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = StableUnCLIPImageNormalizer(embedding_dim=__lowercase ) lowerCamelCase__ : int = DDPMScheduler(beta_schedule="""squaredcos_cap_v2""" ) torch.manual_seed(0 ) lowerCamelCase__ : List[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) torch.manual_seed(0 ) lowerCamelCase__ : int = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) ) torch.manual_seed(0 ) lowerCamelCase__ : Optional[int] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """CrossAttnUpBlock2D""") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="""projection""" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0 ) lowerCamelCase__ : Dict = DDIMScheduler( beta_schedule="""scaled_linear""" , beta_start=0.00_085 , beta_end=0.012 , prediction_type="""v_prediction""" , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0 ) lowerCamelCase__ : Optional[Any] = AutoencoderKL() lowerCamelCase__ : Union[str, Any] = { # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Dict=0 ): if str(__lowercase ).startswith("""mps""" ): lowerCamelCase__ : List[Any] = torch.manual_seed(__lowercase ) else: lowerCamelCase__ : List[Any] = torch.Generator(device=__lowercase ).manual_seed(__lowercase ) lowerCamelCase__ : List[Any] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : List[str] = torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=__lowercase ) def lowerCamelCase_ ( self: Union[str, Any] ): lowerCamelCase__ : str = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase ) @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: str ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self: Optional[int] ): lowerCamelCase__ : Optional[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy""" ) lowerCamelCase__ : Tuple = StableUnCLIPPipeline.from_pretrained("""fusing/stable-unclip-2-1-l""" , torch_dtype=torch.floataa ) pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCamelCase__ : Dict = torch.Generator(device="""cpu""" ).manual_seed(0 ) lowerCamelCase__ : Optional[Any] = pipe("""anime turle""" , generator=__lowercase , output_type="""np""" ) lowerCamelCase__ : Union[str, Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase ) def lowerCamelCase_ ( self: Optional[Any] ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCamelCase__ : List[str] = StableUnCLIPPipeline.from_pretrained("""fusing/stable-unclip-2-1-l""" , torch_dtype=torch.floataa ) lowerCamelCase__ : Dict = pipe.to(__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCamelCase__ : Dict = pipe( """anime turtle""" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="""np""" , ) lowerCamelCase__ : Dict = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
350
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("""The given input must be positive""" ) # get the generated string sequence lowerCamelCase__ : List[str] = gray_code_sequence_string(UpperCamelCase ) # # convert them to integers for i in range(len(UpperCamelCase ) ): lowerCamelCase__ : int = int(sequence[i] , 2 ) return sequence def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCamelCase__ : Optional[Any] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCamelCase__ : Any = gray_code_sequence_string(bit_count - 1 ) lowerCamelCase__ : Union[str, Any] = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCamelCase__ : Optional[int] = """0""" + smaller_sequence[i] sequence.append(UpperCamelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCamelCase__ : int = """1""" + smaller_sequence[i] sequence.append(UpperCamelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
129
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) lowerCAmelCase = {"configuration_vit": ["VIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTConfig", "ViTOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = ["ViTFeatureExtractor"] lowerCAmelCase = ["ViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ "VIT_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTForImageClassification", "ViTForMaskedImageModeling", "ViTModel", "ViTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ "TFViTForImageClassification", "TFViTModel", "TFViTPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ "FlaxViTForImageClassification", "FlaxViTModel", "FlaxViTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys lowerCAmelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
126
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase : List[Any] =logging.get_logger(__name__) def UpperCamelCase ( _lowerCamelCase : Optional[int] , _lowerCamelCase : Union[str, Any]=False ): A__ = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight") ) rename_keys.append((F"patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias", F"vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias") ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((F"blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A__ = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def UpperCamelCase ( _lowerCamelCase : Tuple , _lowerCamelCase : int , _lowerCamelCase : List[Any]=False ): for i in range(config.num_hidden_layers ): if base_model: A__ = "" else: A__ = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) A__ = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict A__ = in_proj_weight[ : config.hidden_size, : ] A__ = in_proj_bias[: config.hidden_size] A__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ = in_proj_weight[ -config.hidden_size :, : ] A__ = in_proj_bias[-config.hidden_size :] def UpperCamelCase ( _lowerCamelCase : int ): A__ = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(_lowerCamelCase , _lowerCamelCase ) def UpperCamelCase ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Tuple , _lowerCamelCase : List[Any] ): A__ = dct.pop(_lowerCamelCase ) A__ = val def UpperCamelCase ( ): A__ = "http://images.cocodataset.org/val2017/000000039769.jpg" A__ = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase ).raw ) return im @torch.no_grad() def UpperCamelCase ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Any , _lowerCamelCase : Optional[Any]=False ): A__ = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=_lowerCamelCase , ) A__ = ViTHybridConfig(backbone_config=_lowerCamelCase , image_size=3_84 , num_labels=10_00 ) A__ = False # load original model from timm A__ = timm.create_model(_lowerCamelCase , pretrained=_lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ = timm_model.state_dict() if base_model: remove_classification_head_(_lowerCamelCase ) A__ = create_rename_keys(_lowerCamelCase , _lowerCamelCase ) for src, dest in rename_keys: rename_key(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) read_in_q_k_v(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A__ = "huggingface/label-files" A__ = "imagenet-1k-id2label.json" A__ = json.load(open(hf_hub_download(_lowerCamelCase , _lowerCamelCase , repo_type="dataset" ) , "r" ) ) A__ = {int(_lowerCamelCase ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": A__ = ViTHybridModel(_lowerCamelCase ).eval() else: A__ = ViTHybridForImageClassification(_lowerCamelCase ).eval() model.load_state_dict(_lowerCamelCase ) # create image processor A__ = create_transform(**resolve_data_config({} , model=_lowerCamelCase ) ) A__ = transform.transforms A__ = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } A__ = ViTHybridImageProcessor( do_resize=_lowerCamelCase , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=_lowerCamelCase , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=_lowerCamelCase , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) A__ = prepare_img() A__ = transform(_lowerCamelCase ).unsqueeze(0 ) A__ = processor(_lowerCamelCase , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(_lowerCamelCase , _lowerCamelCase ) # verify logits with torch.no_grad(): A__ = model(_lowerCamelCase ) A__ = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: A__ = timm_model.forward_features(_lowerCamelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_lowerCamelCase , outputs.pooler_output , atol=1e-3 ) else: A__ = timm_model(_lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCamelCase , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) print(F"Saving model {vit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_lowerCamelCase ) print(F"Saving processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(_lowerCamelCase ) if push_to_hub: print(F"Pushing model and processor to the hub {vit_name}" ) model.push_to_hub(F"ybelkada/{vit_name}" ) processor.push_to_hub(F"ybelkada/{vit_name}" ) if __name__ == "__main__": __lowerCAmelCase : List[str] =argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_r50_s16_384", type=str, help="Name of the hybrid ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to upload the model to the HuggingFace hub." ) __lowerCAmelCase : Optional[int] =parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
237
0
"""simple docstring""" import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): a__ = LayoutLMTokenizer a__ = LayoutLMTokenizerFast a__ = True a__ = True def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' super().setUp() a__: List[str] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] a__: List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file']) with open(self.vocab_file , 'w' , encoding='utf-8') as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens])) def lowerCamelCase_ ( self , **lowercase) -> Union[str, Any]: '''simple docstring''' return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **lowercase) def lowerCamelCase_ ( self , lowercase) -> Tuple: '''simple docstring''' a__: List[str] = 'UNwant\u00E9d,running' a__: Optional[int] = 'unwanted, running' return input_text, output_text def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: str = self.tokenizer_class(self.vocab_file) a__: str = tokenizer.tokenize('UNwant\u00E9d,running') self.assertListEqual(lowercase , ['un', '##want', '##ed', ',', 'runn', '##ing']) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase) , [7, 4, 5, 10, 8, 9]) def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' pass
359
"""simple docstring""" import argparse import math import traceback import dateutil.parser as date_parser import requests def __a ( _SCREAMING_SNAKE_CASE ) ->Tuple: a__: Tuple = {} a__: Tuple = job['started_at'] a__: int = job['completed_at'] a__: Any = date_parser.parse(_SCREAMING_SNAKE_CASE ) a__: Tuple = date_parser.parse(_SCREAMING_SNAKE_CASE ) a__: str = round((end_datetime - start_datetime).total_seconds() / 60.0 ) a__: Any = start a__: Dict = end a__: Optional[int] = duration_in_min return job_info def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Optional[int]: a__: Tuple = None if token is not None: a__: List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} a__: int = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' a__: Union[str, Any] = requests.get(_SCREAMING_SNAKE_CASE , headers=_SCREAMING_SNAKE_CASE ).json() a__: str = {} try: job_time.update({job['name']: extract_time_from_single_job(_SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) a__: Dict = math.ceil((result['total_count'] - 100) / 100 ) for i in range(_SCREAMING_SNAKE_CASE ): a__: str = requests.get(url + F'&page={i + 2}' , headers=_SCREAMING_SNAKE_CASE ).json() job_time.update({job['name']: extract_time_from_single_job(_SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) return job_time except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') lowercase__ = parser.parse_args() lowercase__ = get_job_time(args.workflow_run_id) lowercase__ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f"{k}: {v['duration']}")
203
0
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed lowerCAmelCase : Optional[Any] = """true""" def A_ ( _UpperCAmelCase , _UpperCAmelCase=82 , _UpperCAmelCase=16 ): set_seed(42 ) SCREAMING_SNAKE_CASE_: List[str] = RegressionModel() SCREAMING_SNAKE_CASE_: int = deepcopy(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] = RegressionDataset(length=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Any = DataLoader(_UpperCAmelCase , batch_size=_UpperCAmelCase ) model.to(accelerator.device ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = accelerator.prepare(_UpperCAmelCase , _UpperCAmelCase ) return model, ddp_model, dataloader def A_ ( _UpperCAmelCase , _UpperCAmelCase=False ): SCREAMING_SNAKE_CASE_: Optional[int] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) SCREAMING_SNAKE_CASE_: Any = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[Any] = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase ) return outputs with accelerator.main_process_first(): SCREAMING_SNAKE_CASE_: Optional[Any] = dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) SCREAMING_SNAKE_CASE_: int = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(_UpperCAmelCase ): if use_longest: return tokenizer.pad(_UpperCAmelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(_UpperCAmelCase , padding="max_length" , max_length=1_28 , return_tensors="pt" ) return DataLoader(_UpperCAmelCase , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=16 ) def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Any = Accelerator(dispatch_batches=_UpperCAmelCase , split_batches=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = get_dataloader(_UpperCAmelCase , not dispatch_batches ) SCREAMING_SNAKE_CASE_: Tuple = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = accelerator.prepare(_UpperCAmelCase , _UpperCAmelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Union[str, Any] = [] for batch in dataloader: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = batch.values() with torch.no_grad(): SCREAMING_SNAKE_CASE_: List[Any] = model(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Dict = [], [] for logit, targ in logits_and_targets: logits.append(_UpperCAmelCase ) targs.append(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = torch.cat(_UpperCAmelCase ), torch.cat(_UpperCAmelCase ) return logits, targs def A_ ( _UpperCAmelCase , _UpperCAmelCase=82 , _UpperCAmelCase=False , _UpperCAmelCase=False , _UpperCAmelCase=16 ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = get_basic_setup(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = generate_predictions(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) assert ( len(_UpperCAmelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(_UpperCAmelCase )}" def A_ ( _UpperCAmelCase = False , _UpperCAmelCase = False ): SCREAMING_SNAKE_CASE_: str = evaluate.load("glue" , "mrpc" ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = get_mrpc_setup(_UpperCAmelCase , _UpperCAmelCase ) # First do baseline SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = setup["no"] model.to(_UpperCAmelCase ) model.eval() for batch in dataloader: batch.to(_UpperCAmelCase ) with torch.inference_mode(): SCREAMING_SNAKE_CASE_: str = model(**_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Any = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=_UpperCAmelCase , references=batch["labels"] ) SCREAMING_SNAKE_CASE_: int = metric.compute() # Then do distributed SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): SCREAMING_SNAKE_CASE_: List[str] = model(**_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] = outputs.logits.argmax(dim=-1 ) SCREAMING_SNAKE_CASE_: Union[str, Any] = batch["labels"] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=_UpperCAmelCase , references=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def A_ ( ): SCREAMING_SNAKE_CASE_: Any = Accelerator(split_batches=_UpperCAmelCase , dispatch_batches=_UpperCAmelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(_UpperCAmelCase , _UpperCAmelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: SCREAMING_SNAKE_CASE_: Tuple = Accelerator(split_batches=_UpperCAmelCase , dispatch_batches=_UpperCAmelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(_UpperCAmelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) SCREAMING_SNAKE_CASE_: int = Accelerator() test_torch_metrics(_UpperCAmelCase , 5_12 ) accelerator.state._reset_state() def A_ ( _UpperCAmelCase ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
13
"""simple docstring""" def UpperCAmelCase ( UpperCAmelCase ) -> list: if len(UpperCAmelCase ) <= 1: return [tuple(UpperCAmelCase )] snake_case_ = [] def generate(UpperCAmelCase , UpperCAmelCase ): snake_case_ = [0] * n res.append(tuple(UpperCAmelCase ) ) snake_case_ = 0 while i < n: if c[i] < i: if i % 2 == 0: snake_case_ , snake_case_ = arr[i], arr[0] else: snake_case_ , snake_case_ = arr[i], arr[c[i]] res.append(tuple(UpperCAmelCase ) ) c[i] += 1 snake_case_ = 0 else: snake_case_ = 0 i += 1 generate(len(UpperCAmelCase ) , UpperCAmelCase ) return res if __name__ == "__main__": __UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
69
0
'''simple docstring''' from __future__ import annotations _lowercase : Tuple = tuple[int, int, int] _lowercase : str = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase _lowercase : str = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- _lowercase : Any = "EGZWVONAHDCLFQMSIPJBYUKXTR" _lowercase : Any = "FOBHMDKEXQNRAULPGSJVTYICZW" _lowercase : Dict = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- _lowercase : Dict = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- _lowercase : Union[str, Any] = "RMDJXFUWGISLHVTCQNKYPBEZOA" _lowercase : Dict = "SGLCPQWZHKXAREONTFBVIYJUDM" _lowercase : Union[str, Any] = "HVSICLTYKQUBXDWAJZOMFGPREN" _lowercase : Any = "RZWQHFMVDBKICJLNTUXAGYPSOE" _lowercase : List[Any] = "LFKIJODBEGAMQPXVUHYSTCZRWN" _lowercase : int = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def snake_case_ ( __SCREAMING_SNAKE_CASE : RotorPositionT , __SCREAMING_SNAKE_CASE : RotorSelectionT , __SCREAMING_SNAKE_CASE : str ): """simple docstring""" if (unique_rotsel := len(set(__SCREAMING_SNAKE_CASE ) )) < 3: lowercase_ : Any = F'''Please use 3 unique rotors (not {unique_rotsel})''' raise Exception(__SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid lowercase_ , lowercase_ , lowercase_ : Dict = rotpos if not 0 < rotorposa <= len(__SCREAMING_SNAKE_CASE ): lowercase_ : str = F'''First rotor position is not within range of 1..26 ({rotorposa}''' raise ValueError(__SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(__SCREAMING_SNAKE_CASE ): lowercase_ : Optional[int] = F'''Second rotor position is not within range of 1..26 ({rotorposa})''' raise ValueError(__SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(__SCREAMING_SNAKE_CASE ): lowercase_ : str = F'''Third rotor position is not within range of 1..26 ({rotorposa})''' raise ValueError(__SCREAMING_SNAKE_CASE ) # Validates string and returns dict lowercase_ : str = _plugboard(__SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def snake_case_ ( __SCREAMING_SNAKE_CASE : str ): """simple docstring""" if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : Optional[Any] = F'''Plugboard setting isn\'t type string ({type(__SCREAMING_SNAKE_CASE )})''' raise TypeError(__SCREAMING_SNAKE_CASE ) elif len(__SCREAMING_SNAKE_CASE ) % 2 != 0: lowercase_ : Dict = F'''Odd number of symbols ({len(__SCREAMING_SNAKE_CASE )})''' raise Exception(__SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(''' ''' , '''''' ) # Checks if all characters are unique lowercase_ : Any = set() for i in pbstring: if i not in abc: lowercase_ : Dict = F'''\'{i}\' not in list of symbols''' raise Exception(__SCREAMING_SNAKE_CASE ) elif i in tmppbl: lowercase_ : Any = F'''Duplicate symbol ({i})''' raise Exception(__SCREAMING_SNAKE_CASE ) else: tmppbl.add(__SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary lowercase_ : Dict = {} for j in range(0 , len(__SCREAMING_SNAKE_CASE ) - 1 , 2 ): lowercase_ : str = pbstring[j + 1] lowercase_ : Optional[Any] = pbstring[j] return pb def snake_case_ ( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : RotorPositionT , __SCREAMING_SNAKE_CASE : RotorSelectionT = (rotora, rotora, rotora) , __SCREAMING_SNAKE_CASE : str = "" , ): """simple docstring""" lowercase_ : Optional[int] = text.upper() lowercase_ , lowercase_ , lowercase_ : int = _validator( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , plugb.upper() ) lowercase_ , lowercase_ , lowercase_ : Tuple = rotor_position lowercase_ , lowercase_ , lowercase_ : str = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 lowercase_ : Tuple = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: lowercase_ : Tuple = plugboard[symbol] # rotor ra -------------------------- lowercase_ : str = abc.index(__SCREAMING_SNAKE_CASE ) + rotorposa lowercase_ : str = rotora[index % len(__SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- lowercase_ : Any = abc.index(__SCREAMING_SNAKE_CASE ) + rotorposa lowercase_ : Optional[int] = rotora[index % len(__SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- lowercase_ : str = abc.index(__SCREAMING_SNAKE_CASE ) + rotorposa lowercase_ : Any = rotora[index % len(__SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher lowercase_ : Optional[Any] = reflector[symbol] # 2nd rotors lowercase_ : List[str] = abc[rotora.index(__SCREAMING_SNAKE_CASE ) - rotorposa] lowercase_ : Dict = abc[rotora.index(__SCREAMING_SNAKE_CASE ) - rotorposa] lowercase_ : Optional[int] = abc[rotora.index(__SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: lowercase_ : str = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(__SCREAMING_SNAKE_CASE ): lowercase_ : Dict = 0 rotorposa += 1 if rotorposa >= len(__SCREAMING_SNAKE_CASE ): lowercase_ : str = 0 rotorposa += 1 if rotorposa >= len(__SCREAMING_SNAKE_CASE ): lowercase_ : int = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(__SCREAMING_SNAKE_CASE ) return "".join(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": _lowercase : Tuple = "This is my Python script that emulates the Enigma machine from WWII." _lowercase : Union[str, Any] = (1, 1, 1) _lowercase : Optional[int] = "pictures" _lowercase : List[Any] = (rotora, rotora, rotora) _lowercase : Dict = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
264
'''simple docstring''' from __future__ import annotations from typing import Any class lowerCAmelCase__ : def __init__( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : str = num_of_nodes lowercase_ : list[list[int]] = [] lowercase_ : dict[int, int] = {} def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" self.m_edges.append([u_node, v_node, weight] ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" if self.m_component[u_node] != u_node: for k in self.m_component: lowercase_ : Optional[int] = self.find_component(__SCREAMING_SNAKE_CASE ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" if component_size[u_node] <= component_size[v_node]: lowercase_ : Any = v_node component_size[v_node] += component_size[u_node] self.set_component(__SCREAMING_SNAKE_CASE ) elif component_size[u_node] >= component_size[v_node]: lowercase_ : int = self.find_component(__SCREAMING_SNAKE_CASE ) component_size[u_node] += component_size[v_node] self.set_component(__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : Dict = [] lowercase_ : Optional[Any] = 0 lowercase_ : list[Any] = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) lowercase_ : Union[str, Any] = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: lowercase_ , lowercase_ , lowercase_ : List[Any] = edge lowercase_ : Dict = self.m_component[u] lowercase_ : Any = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): lowercase_ : Union[str, Any] = [u, v, w] for edge in minimum_weight_edge: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ , lowercase_ , lowercase_ : str = edge lowercase_ : Tuple = self.m_component[u] lowercase_ : Union[str, Any] = self.m_component[v] if u_component != v_component: mst_weight += w self.union(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) print(F'''Added edge [{u} - {v}]\nAdded weight: {w}\n''' ) num_of_components -= 1 lowercase_ : str = [-1] * self.m_num_of_nodes print(F'''The total weight of the minimal spanning tree is: {mst_weight}''' ) def snake_case_ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
264
1
from math import asin, atan, cos, radians, sin, sqrt, tan lowerCAmelCase__ :Optional[int] = 6_378_137.0 lowerCAmelCase__ :Dict = 6_356_752.314_245 lowerCAmelCase__ :List[str] = 6_3_7_8_1_3_7 def lowerCAmelCase__ ( a__: float , a__: float , a__: float , a__: float ) -> Optional[int]: '''simple docstring''' _UpperCAmelCase = (AXIS_A - AXIS_B) / AXIS_A _UpperCAmelCase = atan((1 - flattening) * tan(radians(lowerCamelCase_ ) ) ) _UpperCAmelCase = atan((1 - flattening) * tan(radians(lowerCamelCase_ ) ) ) _UpperCAmelCase = radians(lowerCamelCase_ ) _UpperCAmelCase = radians(lowerCamelCase_ ) # Equation _UpperCAmelCase = sin((phi_a - phi_a) / 2 ) _UpperCAmelCase = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda _UpperCAmelCase = sqrt(sin_sq_phi + (cos(lowerCamelCase_ ) * cos(lowerCamelCase_ ) * sin_sq_lambda) ) return 2 * RADIUS * asin(lowerCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
329
import os import jsonlines import numpy as np from tqdm import tqdm __snake_case : Any =2_0_4_8 __snake_case : Union[str, Any] =4_0_9_6 __snake_case : Optional[Any] =4_2 __snake_case : Dict =os.environ.pop('PROCESS_TRAIN', 'false') __snake_case : List[str] ={'null': 0, 'short': 1, 'long': 2, 'yes': 3, 'no': 4} def lowerCAmelCase__ ( lowerCamelCase_ : List[Any]): '''simple docstring''' def choose_first(lowerCamelCase_ : List[str] ,lowerCamelCase_ : Any=False): assert isinstance(lowerCamelCase_ ,lowerCamelCase_) if len(lowerCamelCase_) == 1: lowerCAmelCase__ : Optional[int] = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: lowerCAmelCase__ : Any = {k: [a[k]] for k in a} if len(a['''start_token''']) > 0: break return a lowerCAmelCase__ : Optional[Any] = {'''id''': example['''id''']} lowerCAmelCase__ : int = example['''annotations'''] lowerCAmelCase__ : str = annotation['''yes_no_answer'''] if 0 in yes_no_answer or 1 in yes_no_answer: lowerCAmelCase__ : Union[str, Any] = ['''yes'''] if 1 in yes_no_answer else ['''no'''] lowerCAmelCase__ : int = [] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : int = ['''<cls>'''] else: lowerCAmelCase__ : Tuple = ['''short'''] lowerCAmelCase__ : int = choose_first(annotation['''short_answers''']) if len(out['''start_token''']) == 0: # answer will be long if short is not available lowerCAmelCase__ : Optional[Any] = ['''long'''] lowerCAmelCase__ : str = choose_first(annotation['''long_answer'''] ,is_long_answer=lowerCamelCase_) lowerCAmelCase__ : Optional[int] = [] answer.update(lowerCamelCase_) # disregard some samples if len(answer['''start_token''']) > 1 or answer["start_token"] == answer["end_token"]: lowerCAmelCase__ : Optional[Any] = True else: lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : Tuple = ['''start_token''', '''end_token''', '''start_byte''', '''end_byte''', '''text'''] if not all(isinstance(answer[k] ,lowerCamelCase_) for k in cols): raise ValueError('''Issue in ID''' ,example['''id''']) return answer def lowerCAmelCase__ ( lowerCamelCase_ : List[Any] ,lowerCamelCase_ : Union[str, Any]=False): '''simple docstring''' lowerCAmelCase__ : Any = _get_single_answer(lowerCamelCase_) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase__ : List[Any] = example['''document''']['''tokens'''] lowerCAmelCase__ : Any = [] for i in range(len(doc['''token'''])): if not doc["is_html"][i]: context.append(doc['''token'''][i]) return { "context": " ".join(lowerCamelCase_), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples lowerCAmelCase__ : Union[str, Any] = ['''start_token''', '''end_token'''] answer.update({k: answer[k][0] if len(answer[k]) > 0 else answer[k] for k in cols}) # e.g. [10] == 10 lowerCAmelCase__ : List[Any] = example['''document''']['''tokens'''] lowerCAmelCase__ : Optional[Any] = answer['''start_token'''] lowerCAmelCase__ : Union[str, Any] = answer['''end_token'''] lowerCAmelCase__ : int = [] for i in range(len(doc['''token'''])): if not doc["is_html"][i]: context.append(doc['''token'''][i]) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 lowerCAmelCase__ : List[Any] = ''' '''.join(context[start_token:end_token]) # checking above code if assertion: lowerCAmelCase__ : str = doc['''is_html'''][answer['''start_token'''] : answer['''end_token''']] lowerCAmelCase__ : List[Any] = doc['''token'''][answer['''start_token'''] : answer['''end_token''']] lowerCAmelCase__ : Optional[int] = ''' '''.join([old[i] for i in range(len(lowerCamelCase_)) if not is_html[i]]) if new != old: print('''ID:''' ,example['''id''']) print('''New:''' ,lowerCamelCase_ ,end='''\n''') print('''Old:''' ,lowerCamelCase_ ,end='''\n\n''') return { "context": " ".join(lowerCamelCase_), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def lowerCAmelCase__ ( lowerCamelCase_ : int ,lowerCamelCase_ : str ,lowerCamelCase_ : Tuple=2048 ,lowerCamelCase_ : Dict=4096 ,lowerCamelCase_ : Optional[Any]=True): '''simple docstring''' lowerCAmelCase__ : int = get_context_and_ans(lowerCamelCase_ ,assertion=lowerCamelCase_) lowerCAmelCase__ : Union[str, Any] = out['''answer'''] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } lowerCAmelCase__ : Union[str, Any] = tokenizer(example['''question''']['''text'''] ,out['''context''']).input_ids lowerCAmelCase__ : List[str] = input_ids.index(tokenizer.sep_token_id) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : Dict = [] lowerCAmelCase__ : List[Any] = input_ids[:q_len] lowerCAmelCase__ : List[Any] = range(lowerCamelCase_ ,len(lowerCamelCase_) ,max_length - doc_stride) for i in doc_start_indices: lowerCAmelCase__ : Union[str, Any] = i + max_length - q_len lowerCAmelCase__ : Any = input_ids[i:end_index] inputs.append(q_indices + slice) category.append(answer['''category'''][0]) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(lowerCamelCase_), "end_token": [-100] * len(lowerCamelCase_), "category": category, }, } lowerCAmelCase__ : Optional[Any] = out['''context'''].split() lowerCAmelCase__ : Union[str, Any] = splitted_context[answer['''end_token''']] lowerCAmelCase__ : Optional[int] = len( tokenizer( ''' '''.join(splitted_context[: answer['''start_token''']]) ,add_special_tokens=lowerCamelCase_ ,).input_ids) lowerCAmelCase__ : Dict = len( tokenizer(''' '''.join(splitted_context[: answer['''end_token''']]) ,add_special_tokens=lowerCamelCase_).input_ids) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token lowerCAmelCase__ : int = len(tokenizer(lowerCamelCase_ ,add_special_tokens=lowerCamelCase_).input_ids) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 lowerCAmelCase__ : Union[str, Any] = input_ids[answer['''start_token'''] : answer['''end_token'''] + 1] # right & left are inclusive lowerCAmelCase__ : List[str] = answer['''start_token'''] lowerCAmelCase__ : Union[str, Any] = answer['''end_token'''] if assertion: lowerCAmelCase__ : int = tokenizer.decode(lowerCamelCase_) if answer["span"] != new: print('''ISSUE IN TOKENIZATION''') print('''OLD:''' ,answer['''span''']) print('''NEW:''' ,lowerCamelCase_ ,end='''\n\n''') if len(lowerCamelCase_) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } lowerCAmelCase__ : int = input_ids[:q_len] lowerCAmelCase__ : Optional[Any] = range(lowerCamelCase_ ,len(lowerCamelCase_) ,max_length - doc_stride) lowerCAmelCase__ : Tuple = [] lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : Any = [] # null, yes, no, long, short for i in doc_start_indices: lowerCAmelCase__ : str = i + max_length - q_len lowerCAmelCase__ : List[str] = input_ids[i:end_index] inputs.append(q_indices + slice) assert len(inputs[-1]) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: lowerCAmelCase__ : int = start_token - i + q_len lowerCAmelCase__ : str = end_token - i + q_len answers_category.append(answer['''category'''][0]) # ["short"] -> "short" else: lowerCAmelCase__ : Tuple = -100 lowerCAmelCase__ : List[str] = -100 answers_category.append('''null''') lowerCAmelCase__ : int = inputs[-1][start_token : end_token + 1] answers_start_token.append(lowerCamelCase_) answers_end_token.append(lowerCamelCase_) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('''ISSUE in strided for ID:''' ,example['''id''']) print('''New:''' ,tokenizer.decode(lowerCamelCase_)) print('''Old:''' ,tokenizer.decode(lowerCamelCase_) ,end='''\n\n''') if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def lowerCAmelCase__ ( lowerCamelCase_ : Any ,lowerCamelCase_ : List[Any] ,lowerCamelCase_ : int=2048 ,lowerCamelCase_ : Tuple=4096 ,lowerCamelCase_ : Optional[int]=False): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = get_strided_contexts_and_ans( lowerCamelCase_ ,lowerCamelCase_ ,doc_stride=lowerCamelCase_ ,max_length=lowerCamelCase_ ,assertion=lowerCamelCase_ ,) return example def lowerCAmelCase__ ( lowerCamelCase_ : str ,lowerCamelCase_ : int): '''simple docstring''' with jsonlines.open(lowerCamelCase_ ,'''a''') as writer: for example in tqdm(lowerCamelCase_ ,total=len(lowerCamelCase_) ,desc='''Saving samples ... '''): lowerCAmelCase__ : Optional[Any] = example['''labels'''] for ids, start, end, cat in zip( example['''input_ids'''] ,labels['''start_token'''] ,labels['''end_token'''] ,labels['''category'''] ,): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { '''input_ids''': ids, '''start_token''': start, '''end_token''': end, '''category''': CATEGORY_MAPPING[cat], }) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer __snake_case : Optional[int] =load_dataset('natural_questions') __snake_case : Union[str, Any] =BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base') __snake_case : Tuple =data['train' if PROCESS_TRAIN == 'true' else 'validation'] __snake_case : Optional[int] ={ 'tokenizer': tokenizer, 'doc_stride': DOC_STRIDE, 'max_length': MAX_LENGTH, 'assertion': False, } __snake_case : Dict =data.map(prepare_inputs, fn_kwargs=fn_kwargs) __snake_case : Dict =data.remove_columns(['annotations', 'document', 'id', 'question']) print(data) np.random.seed(SEED) __snake_case : int ='nq-training.jsonl' if PROCESS_TRAIN == 'true' else 'nq-validation.jsonl' save_to_disk(data, file_name=cache_file_name)
129
0
"""simple docstring""" import numpy as np from PIL import Image def __lowercase ( a__ , a__ , a__ ) -> np.ndarray: __SCREAMING_SNAKE_CASE = np.array(a__ ) if arr.shape[0] != arr.shape[1]: raise ValueError('The input array is not a square matrix' ) __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 # compute the shape of the output matrix __SCREAMING_SNAKE_CASE = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __SCREAMING_SNAKE_CASE = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __SCREAMING_SNAKE_CASE = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 return updated_arr def __lowercase ( a__ , a__ , a__ ) -> np.ndarray: __SCREAMING_SNAKE_CASE = np.array(a__ ) if arr.shape[0] != arr.shape[1]: raise ValueError('The input array is not a square matrix' ) __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 # compute the shape of the output matrix __SCREAMING_SNAKE_CASE = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __SCREAMING_SNAKE_CASE = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __SCREAMING_SNAKE_CASE = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name='''avgpooling''', verbose=True) # Loading the image lowerCAmelCase__ : int =Image.open('''path_to_image''') # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
366
from collections import deque from .hash_table import HashTable class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' def __init__( self , *_A , **_A ): '''simple docstring''' super().__init__(*_A , **_A ) def _A ( self , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_A ) __SCREAMING_SNAKE_CASE = self.values[key] def _A ( self ): '''simple docstring''' return ( sum(self.charge_factor - len(_A ) for slot in self.values ) / self.size_table * self.charge_factor ) def _A ( self , _A , _A=None ): '''simple docstring''' if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_A ) == 0 ): return key return super()._collision_resolution(_A , _A )
118
0