code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _A ( lowerCAmelCase_ : List[Any] ): """simple docstring""" lowerCAmelCase__ = filter(lambda lowerCAmelCase_ : p.requires_grad , model.parameters() ) lowerCAmelCase__ = sum([np.prod(p.size() ) for p in model_parameters] ) return params UpperCamelCase = logging.getLogger(__name__) def _A ( lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Any ): """simple docstring""" if metric == "rouge2": lowerCAmelCase__ = "{val_avg_rouge2:.4f}-{step_count}" elif metric == "bleu": lowerCAmelCase__ = "{val_avg_bleu:.4f}-{step_count}" elif metric == "em": lowerCAmelCase__ = "{val_avg_em:.4f}-{step_count}" elif metric == "loss": lowerCAmelCase__ = "{val_avg_loss:.4f}-{step_count}" else: raise NotImplementedError( F'seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this' " function." ) lowerCAmelCase__ = ModelCheckpoint( dirpath=lowerCAmelCase_ , filename=lowerCAmelCase_ , monitor=F'val_{metric}' , mode="max" , save_top_k=1 , every_n_epochs=1 , ) return checkpoint_callback def _A ( lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Dict ): """simple docstring""" return EarlyStopping( monitor=F'val_{metric}' , mode="min" if "loss" in metric else "max" , patience=lowerCAmelCase_ , verbose=lowerCAmelCase_ , ) class __lowerCamelCase ( pl.Callback ): """simple docstring""" def a ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> Dict: lowerCAmelCase__ = {f'lr_group_{i}': param["lr"] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(SCREAMING_SNAKE_CASE__ ) @rank_zero_only def a ( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : pl.Trainer , SCREAMING_SNAKE_CASE__ : pl.LightningModule , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : List[Any]=True ) -> None: logger.info(f'***** {type_path} results at step {trainer.global_step:05d} *****' ) lowerCAmelCase__ = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results lowerCAmelCase__ = Path(pl_module.hparams.output_dir ) if type_path == "test": lowerCAmelCase__ = od / "test_results.txt" lowerCAmelCase__ = od / "test_generations.txt" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. lowerCAmelCase__ = od / f'{type_path}_results/{trainer.global_step:05d}.txt' lowerCAmelCase__ = od / f'{type_path}_generations/{trainer.global_step:05d}.txt' results_file.parent.mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) generations_file.parent.mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) with open(SCREAMING_SNAKE_CASE__ , "a+" ) as writer: for key in sorted(SCREAMING_SNAKE_CASE__ ): if key in ["log", "progress_bar", "preds"]: continue lowerCAmelCase__ = metrics[key] if isinstance(SCREAMING_SNAKE_CASE__ , torch.Tensor ): lowerCAmelCase__ = val.item() lowerCAmelCase__ = f'{key}: {val:.6f}\n' writer.write(SCREAMING_SNAKE_CASE__ ) if not save_generations: return if "preds" in metrics: lowerCAmelCase__ = "\n".join(metrics["preds"] ) generations_file.open("w+" ).write(SCREAMING_SNAKE_CASE__ ) @rank_zero_only def a ( self : List[Any] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> List[Any]: try: lowerCAmelCase__ = pl_module.model.model.num_parameters() except AttributeError: lowerCAmelCase__ = pl_module.model.num_parameters() lowerCAmelCase__ = count_trainable_parameters(SCREAMING_SNAKE_CASE__ ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1e6, "grad_mp": n_trainable_pars / 1e6} ) @rank_zero_only def a ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : pl.Trainer , SCREAMING_SNAKE_CASE__ : pl.LightningModule ) -> List[str]: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , "test" ) @rank_zero_only def a ( self : Dict , SCREAMING_SNAKE_CASE__ : pl.Trainer , SCREAMING_SNAKE_CASE__ : Dict ) -> Optional[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
61
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''roformer''' def __init__( self , _lowercase=5_0_0_0_0 , _lowercase=None , _lowercase=7_6_8 , _lowercase=1_2 , _lowercase=1_2 , _lowercase=3_0_7_2 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=1_5_3_6 , _lowercase=2 , _lowercase=0.02 , _lowercase=1E-12 , _lowercase=0 , _lowercase=False , _lowercase=True , **_lowercase , ) -> int: '''simple docstring''' super().__init__(pad_token_id=_lowercase , **_lowercase ) snake_case_ : str = vocab_size snake_case_ : Any = hidden_size if embedding_size is None else embedding_size snake_case_ : List[str] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : Tuple = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : List[str] = rotary_value snake_case_ : str = use_cache class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ : Any = {0: """batch""", 1: """sequence"""} snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
58
0
import inspect from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel, VQModel from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' def __init__( self : int , UpperCAmelCase_ : VQModel , UpperCAmelCase_ : UNetaDModel , UpperCAmelCase_ : DDIMScheduler ): super().__init__() self.register_modules(vqvae=UpperCAmelCase_ , unet=UpperCAmelCase_ , scheduler=UpperCAmelCase_ ) @torch.no_grad() def __call__( self : Optional[Any] , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase_ : float = 0.0 , UpperCAmelCase_ : int = 50 , UpperCAmelCase_ : Optional[str] = "pil" , UpperCAmelCase_ : bool = True , **UpperCAmelCase_ : Tuple , ): SCREAMING_SNAKE_CASE : str = randn_tensor( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , generator=UpperCAmelCase_ , ) SCREAMING_SNAKE_CASE : int = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler SCREAMING_SNAKE_CASE : Tuple = latents * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(UpperCAmelCase_ ) # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature SCREAMING_SNAKE_CASE : Dict = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) SCREAMING_SNAKE_CASE : List[Any] = {} if accepts_eta: SCREAMING_SNAKE_CASE : Optional[int] = eta for t in self.progress_bar(self.scheduler.timesteps ): SCREAMING_SNAKE_CASE : Optional[Any] = self.scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) # predict the noise residual SCREAMING_SNAKE_CASE : str = self.unet(UpperCAmelCase_ , UpperCAmelCase_ ).sample # compute the previous noisy sample x_t -> x_t-1 SCREAMING_SNAKE_CASE : int = self.scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ).prev_sample # decode the image latents with the VAE SCREAMING_SNAKE_CASE : int = self.vqvae.decode(UpperCAmelCase_ ).sample SCREAMING_SNAKE_CASE : int = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE : Optional[int] = self.numpy_to_pil(UpperCAmelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase_ )
62
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __lowerCAmelCase ( __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Dict = checkpoints.load_tax_checkpoint(__UpperCamelCase ) snake_case_ : Tuple = flatten_dict(__UpperCamelCase ) return flax_params def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] ): '''simple docstring''' snake_case_ : Optional[Any] = {} snake_case_ : List[Any] = { """token_embedder""": """embeddings""", """encoder_norm""": """layernorm""", """kernel""": """weight""", """.out""": """.output""", """scale""": """weight""", """embedders_0.pos_embedding""": """row_embedder.weight""", """embedders_1.pos_embedding""": """column_embedder.weight""", } snake_case_ : Optional[Any] = { """query""": """attention.query""", """key""": """attention.key""", """value""": """attention.value""", """output.dense""": """output""", """encoder_decoder_attention.o""": """encoder_decoder_attention.attention.o""", """pre_self_attention_layer_norm""": """self_attention.layer_norm""", """pre_cross_attention_layer_norm""": """encoder_decoder_attention.layer_norm""", """mlp.""": """mlp.DenseReluDense.""", """pre_mlp_layer_norm""": """mlp.layer_norm""", """self_attention.o""": """self_attention.attention.o""", """decoder.embeddings.embedding""": """decoder.embed_tokens.weight""", """decoder.relpos_bias.rel_embedding""": """decoder.layer.0.self_attention.attention.relative_attention_bias.weight""", """decoder.decoder_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.logits_dense.weight""": """decoder.lm_head.weight""", } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key snake_case_ : List[Any] = """.""".join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): snake_case_ : List[str] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): snake_case_ : Optional[int] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number snake_case_ : Optional[Any] = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Union[str, Any] = new_key.replace("""encoder""" , """encoder.encoder""" ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number snake_case_ : int = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Dict = flax_dict[key] snake_case_ : Tuple = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): snake_case_ : Optional[int] = torch.from_numpy(converted_dict[key].T ) else: snake_case_ : List[Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : List[str]=False ): '''simple docstring''' snake_case_ : Optional[int] = get_flax_param(__UpperCamelCase ) if not use_large: snake_case_ : Optional[int] = PixaStructVisionConfig() snake_case_ : Optional[Any] = PixaStructTextConfig() else: snake_case_ : Tuple = PixaStructVisionConfig( hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_attention_heads=2_4 , num_hidden_layers=1_8 ) snake_case_ : List[str] = PixaStructTextConfig(hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_heads=2_4 , num_layers=1_8 ) snake_case_ : str = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=__UpperCamelCase ) snake_case_ : Optional[int] = PixaStructForConditionalGeneration(__UpperCamelCase ) snake_case_ : str = rename_and_convert_flax_params(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) snake_case_ : Optional[int] = AutoTokenizer.from_pretrained("""ybelkada/test-pix2struct-tokenizer""" ) snake_case_ : int = PixaStructImageProcessor() snake_case_ : str = PixaStructProcessor(image_processor=__UpperCamelCase , tokenizer=__UpperCamelCase ) if use_large: snake_case_ : Optional[Any] = 4_0_9_6 snake_case_ : int = True # mkdir if needed os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) processor.save_pretrained(__UpperCamelCase ) print("""Model saved in {}""".format(__UpperCamelCase ) ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') __lowerCAmelCase : List[Any] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
58
0
import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class a : """simple docstring""" def __init__( self : Any , __lowercase : Dict , __lowercase : int = 13 , __lowercase : int = 64 , __lowercase : int = 2 , __lowercase : int = 3 , __lowercase : int = 3 , __lowercase : bool = True , __lowercase : bool = True , __lowercase : int = 128 , __lowercase : List[str]=[16, 32, 64, 128] , __lowercase : int = 7 , __lowercase : int = 4 , __lowercase : int = 37 , __lowercase : str = "gelu" , __lowercase : float = 0.1 , __lowercase : float = 0.1 , __lowercase : int = 10 , __lowercase : float = 0.02 , __lowercase : int = 2 , __lowercase : int = 1 , __lowercase : int = 128 , __lowercase : List[int] = [2, 2, 2, 2] , __lowercase : int = 2 , __lowercase : int = 2 , ) -> int: __UpperCAmelCase : Optional[int] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : int = image_size __UpperCAmelCase : Optional[int] = patch_size __UpperCAmelCase : Any = num_channels __UpperCAmelCase : List[str] = is_training __UpperCAmelCase : Optional[Any] = use_labels __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[int] = num_hidden_layers __UpperCAmelCase : Union[str, Any] = num_attention_heads __UpperCAmelCase : Union[str, Any] = intermediate_size __UpperCAmelCase : List[Any] = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : Union[str, Any] = type_sequence_label_size __UpperCAmelCase : Dict = initializer_range __UpperCAmelCase : str = encoder_stride __UpperCAmelCase : List[Any] = num_attention_outputs __UpperCAmelCase : str = embed_dim __UpperCAmelCase : Tuple = embed_dim + 1 __UpperCAmelCase : Tuple = resolution __UpperCAmelCase : Union[str, Any] = depths __UpperCAmelCase : List[Any] = hidden_sizes __UpperCAmelCase : Optional[Any] = dim __UpperCAmelCase : Union[str, Any] = mlp_expansion_ratio def UpperCAmelCase ( self : str ) -> int: __UpperCAmelCase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase : int = None if self.use_labels: __UpperCAmelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase : Any = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self : List[Any] ) -> Optional[int]: return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowercase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def UpperCAmelCase ( self : Optional[int] , __lowercase : int , __lowercase : List[str] , __lowercase : Optional[Any] ) -> Tuple: __UpperCAmelCase : Tuple = TFEfficientFormerModel(config=__lowercase ) __UpperCAmelCase : int = model(__lowercase , training=__lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase ( self : Union[str, Any] , __lowercase : Optional[Any] , __lowercase : Dict , __lowercase : Any ) -> Optional[int]: __UpperCAmelCase : Tuple = self.type_sequence_label_size __UpperCAmelCase : Union[str, Any] = TFEfficientFormerForImageClassification(__lowercase ) __UpperCAmelCase : List[str] = model(__lowercase , labels=__lowercase , training=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __UpperCAmelCase : Dict = 1 __UpperCAmelCase : Optional[Any] = TFEfficientFormerForImageClassification(__lowercase ) __UpperCAmelCase : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __UpperCAmelCase : str = model(__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase ( self : Optional[Any] ) -> int: __UpperCAmelCase : Tuple = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : Union[str, Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class a ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" a : Dict = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) a : int = ( { 'feature-extraction': TFEfficientFormerModel, 'image-classification': ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) a : str = False a : Union[str, Any] = False a : Union[str, Any] = False a : str = False a : int = False def UpperCAmelCase ( self : Tuple ) -> Optional[Any]: __UpperCAmelCase : int = TFEfficientFormerModelTester(self ) __UpperCAmelCase : Optional[int] = ConfigTester( self , config_class=__lowercase , has_text_modality=__lowercase , hidden_size=37 ) def UpperCAmelCase ( self : Optional[int] ) -> Any: self.config_tester.run_common_tests() @unittest.skip(reason="""EfficientFormer does not use inputs_embeds""" ) def UpperCAmelCase ( self : Any ) -> str: pass @unittest.skip(reason="""EfficientFormer does not support input and output embeddings""" ) def UpperCAmelCase ( self : Dict ) -> Dict: pass def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase , __UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : List[Any] = model_class(__lowercase ) __UpperCAmelCase : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : Optional[int] = [*signature.parameters.keys()] __UpperCAmelCase : Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowercase ) def UpperCAmelCase ( self : List[Any] ) -> str: def check_hidden_states_output(__lowercase : Optional[int] , __lowercase : Union[str, Any] , __lowercase : Optional[int] ): __UpperCAmelCase : Optional[int] = model_class(__lowercase ) __UpperCAmelCase : Any = model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) __UpperCAmelCase : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : int = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__lowercase ) , __lowercase ) if hasattr(self.model_tester , """encoder_seq_length""" ): __UpperCAmelCase : List[str] = self.model_tester.encoder_seq_length if hasattr(self.model_tester , """chunk_length""" ) and self.model_tester.chunk_length > 1: __UpperCAmelCase : Any = seq_length * self.model_tester.chunk_length else: __UpperCAmelCase : str = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: __UpperCAmelCase : str = outputs.decoder_hidden_states self.asseretIsInstance(__lowercase , (list, tuple) ) self.assertEqual(len(__lowercase ) , __lowercase ) __UpperCAmelCase : Dict = getattr(self.model_tester , """seq_length""" , __lowercase ) __UpperCAmelCase : str = getattr(self.model_tester , """decoder_seq_length""" , __lowercase ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : Optional[Any] = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Optional[int] = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) def UpperCAmelCase ( self : str , __lowercase : str , __lowercase : Optional[int] , __lowercase : Union[str, Any]=False ) -> Dict: __UpperCAmelCase : Optional[Any] = super()._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase ( self : str ) -> Optional[int]: __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) @unittest.skip(reason="""EfficientFormer does not implement masked image modeling yet""" ) def UpperCAmelCase ( self : str ) -> Any: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowercase ) def UpperCAmelCase ( self : Any ) -> str: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) @slow def UpperCAmelCase ( self : int ) -> int: for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase : Union[str, Any] = TFEfficientFormerModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: __UpperCAmelCase , __UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase : Tuple = True __UpperCAmelCase : Optional[Any] = getattr(self.model_tester , """seq_length""" , __lowercase ) __UpperCAmelCase : Optional[Any] = getattr(self.model_tester , """encoder_seq_length""" , __lowercase ) __UpperCAmelCase : Any = getattr(self.model_tester , """key_length""" , __lowercase ) __UpperCAmelCase : Optional[Any] = getattr(self.model_tester , """chunk_length""" , __lowercase ) if chunk_length is not None and hasattr(self.model_tester , """num_hashes""" ): __UpperCAmelCase : Dict = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: __UpperCAmelCase : List[str] = True __UpperCAmelCase : List[Any] = False __UpperCAmelCase : Dict = True __UpperCAmelCase : Dict = model_class(__lowercase ) __UpperCAmelCase : Optional[Any] = model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) __UpperCAmelCase : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : Any = model_class(__lowercase ) __UpperCAmelCase : str = model(**self._prepare_for_class(__lowercase , __lowercase ) , training=__lowercase ) __UpperCAmelCase : Tuple = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(__lowercase ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def UpperCAmelCase ( self : List[Any] ) -> Any: # We use a simplified version of this test for EfficientFormer because it requires training=False # and Keras refuses to let us force that during functional construction __UpperCAmelCase , __UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model __UpperCAmelCase : int = model_class(__lowercase ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes __UpperCAmelCase : Tuple = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=__lowercase ) for key, val in model.input_signature.items() if key in model.dummy_inputs } __UpperCAmelCase : str = model(__lowercase ) self.assertTrue(outputs_dict is not None ) def lowerCamelCase__ ( ): __UpperCAmelCase : str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class a ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase ( self : Any ) -> List[Any]: return ( EfficientFormerImageProcessor.from_pretrained("""snap-research/efficientformer-l1-300""" ) if is_vision_available() else None ) @slow def UpperCAmelCase ( self : Optional[int] ) -> Optional[int]: __UpperCAmelCase : Any = TFEfficientFormerForImageClassification.from_pretrained("""snap-research/efficientformer-l1-300""" ) __UpperCAmelCase : Optional[Any] = self.default_image_processor __UpperCAmelCase : Optional[Any] = prepare_img() __UpperCAmelCase : Optional[Any] = image_processor(images=__lowercase , return_tensors="""tf""" ) # forward pass __UpperCAmelCase : Optional[Any] = model(**__lowercase , training=__lowercase ) # verify the logits __UpperCAmelCase : Any = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , __lowercase ) __UpperCAmelCase : Optional[int] = tf.constant([-0.0_555, 0.4_825, -0.0_852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowercase , atol=1e-4 ) ) @slow def UpperCAmelCase ( self : Tuple ) -> Dict: __UpperCAmelCase : Tuple = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( """snap-research/efficientformer-l1-300""" ) __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : int = prepare_img() __UpperCAmelCase : Dict = image_processor(images=__lowercase , return_tensors="""tf""" ) # forward pass __UpperCAmelCase : Any = model(**__lowercase , training=__lowercase ) # verify the logits __UpperCAmelCase : Any = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , __lowercase ) __UpperCAmelCase : List[str] = tf.constant([-0.1_312, 0.4_353, -1.0_499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __lowercase , atol=1e-4 ) )
63
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if mass < 0: raise ValueError("""The mass of a body cannot be negative""" ) return 0.5 * mass * abs(__UpperCamelCase ) * abs(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
58
0
import string from math import logaa def A__ ( snake_case_ : str , snake_case_ : str ): SCREAMING_SNAKE_CASE__: Optional[int]= document.translate( str.maketrans('''''' , '''''' , string.punctuation ) ).replace('''\n''' , '''''' ) SCREAMING_SNAKE_CASE__: Tuple= document_without_punctuation.split(''' ''' ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def A__ ( snake_case_ : str , snake_case_ : str ): SCREAMING_SNAKE_CASE__: str= corpus.lower().translate( str.maketrans('''''' , '''''' , string.punctuation ) ) # strip all punctuation and replace it with '' SCREAMING_SNAKE_CASE__: Optional[Any]= corpus_without_punctuation.split('''\n''' ) SCREAMING_SNAKE_CASE__: List[Any]= term.lower() return (len([doc for doc in docs if term in doc] ), len(snake_case_ )) def A__ ( snake_case_ : int , snake_case_ : int , snake_case_ : Tuple=False ): if smoothing: if n == 0: raise ValueError('''log10(0) is undefined.''' ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError('''df must be > 0''' ) elif n == 0: raise ValueError('''log10(0) is undefined.''' ) return round(logaa(n / df ) , 3 ) def A__ ( snake_case_ : int , snake_case_ : int ): return round(tf * idf , 3 )
64
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = StableDiffusionInpaintPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase = frozenset([] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=_lowercase , ) snake_case_ : Dict = PNDMScheduler(skip_prk_steps=_lowercase ) torch.manual_seed(0 ) snake_case_ : str = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) snake_case_ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) snake_case_ : Dict = CLIPTextModel(_lowercase ) snake_case_ : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case_ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self , _lowercase , _lowercase=0 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowercase ) ).to(_lowercase ) snake_case_ : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case_ : Tuple = Image.fromarray(np.uinta(_lowercase ) ).convert("""RGB""" ).resize((6_4, 6_4) ) snake_case_ : Any = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((6_4, 6_4) ) if str(_lowercase ).startswith("""mps""" ): snake_case_ : str = torch.manual_seed(_lowercase ) else: snake_case_ : List[str] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) snake_case_ : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case_ : List[str] = self.get_dummy_components() snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline(**_lowercase ) snake_case_ : Dict = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) snake_case_ : Optional[int] = self.get_dummy_inputs(_lowercase ) snake_case_ : List[str] = sd_pipe(**_lowercase ).images snake_case_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) snake_case_ : Optional[int] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained(_lowercase , safety_checker=_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : List[Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[str] = torch.manual_seed(0 ) snake_case_ : Dict = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ : str = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , torch_dtype=torch.floataa , safety_checker=_lowercase , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : Union[str, Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : Optional[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : str = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case_ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : int = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : List[str] = PNDMScheduler.from_pretrained(_lowercase , subfolder="""scheduler""" ) snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , safety_checker=_lowercase , scheduler=_lowercase , torch_dtype=torch.floataa , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case_ : Any = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , num_inference_steps=2 , output_type="""np""" , ) snake_case_ : Dict = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
58
0
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class __lowercase : def __init__( self : Any ,A : Union[str, Any] ,A : Optional[Any]=2 ,A : Tuple=True ,A : Tuple=False ,A : Optional[int]=10 ,A : Any=3 ,A : Tuple=32 * 8 ,A : List[Any]=32 * 8 ,A : int=4 ,A : List[Any]=64 ,): '''simple docstring''' UpperCAmelCase__ : Tuple = parent UpperCAmelCase__ : Union[str, Any] = batch_size UpperCAmelCase__ : Tuple = is_training UpperCAmelCase__ : Optional[Any] = use_auxiliary_loss UpperCAmelCase__ : int = num_queries UpperCAmelCase__ : Tuple = num_channels UpperCAmelCase__ : List[str] = min_size UpperCAmelCase__ : Optional[Any] = max_size UpperCAmelCase__ : Tuple = num_labels UpperCAmelCase__ : List[str] = hidden_dim UpperCAmelCase__ : Union[str, Any] = hidden_dim def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : str = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( A ) UpperCAmelCase__ : Optional[int] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=A ) UpperCAmelCase__ : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=A ) > 0.5 ).float() UpperCAmelCase__ : Optional[Any] = (torch.rand((self.batch_size, self.num_labels) ,device=A ) > 0.5).long() UpperCAmelCase__ : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : List[Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = MaskaFormerConfig( hidden_size=self.hidden_dim ,) UpperCAmelCase__ : int = self.num_queries UpperCAmelCase__ : str = self.num_labels UpperCAmelCase__ : List[Any] = [1, 1, 1, 1] UpperCAmelCase__ : List[Any] = self.num_channels UpperCAmelCase__ : List[Any] = 64 UpperCAmelCase__ : str = 128 UpperCAmelCase__ : int = self.hidden_dim UpperCAmelCase__ : List[Any] = self.hidden_dim UpperCAmelCase__ : int = self.hidden_dim return config def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = self.prepare_config_and_inputs() UpperCAmelCase__ : List[Any] = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def __lowercase ( self : Optional[int] ,A : Any ,A : int ): '''simple docstring''' UpperCAmelCase__ : int = output.encoder_hidden_states UpperCAmelCase__ : List[str] = output.pixel_decoder_hidden_states UpperCAmelCase__ : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(A ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(A ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(A ) ,config.decoder_layers ) def __lowercase ( self : List[Any] ,A : List[Any] ,A : Dict ,A : Union[str, Any] ,A : str=False ): '''simple docstring''' with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = MaskaFormerModel(config=A ) model.to(A ) model.eval() UpperCAmelCase__ : Tuple = model(pixel_values=A ,pixel_mask=A ) UpperCAmelCase__ : Optional[int] = model(A ,output_hidden_states=A ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(A ,A ) def __lowercase ( self : Tuple ,A : List[str] ,A : Dict ,A : Tuple ,A : Any ,A : Any ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaskaFormerForUniversalSegmentation(config=A ) model.to(A ) model.eval() def comm_check_on_output(A : str ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = model(pixel_values=A ,pixel_mask=A ) UpperCAmelCase__ : Tuple = model(A ) comm_check_on_output(A ) UpperCAmelCase__ : Optional[Any] = model( pixel_values=A ,pixel_mask=A ,mask_labels=A ,class_labels=A ) comm_check_on_output(A ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class __lowercase ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): snake_case_ = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () snake_case_ = {"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def __lowercase ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaskaFormerModelTester(self ) UpperCAmelCase__ : Union[str, Any] = ConfigTester(self ,config_class=A ,has_text_modality=A ) def __lowercase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(A ,**A ,output_hidden_states=A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*A ) @unittest.skip(reason="""Mask2Former does not use inputs_embeds""" ) def __lowercase ( self : List[str] ): '''simple docstring''' pass @unittest.skip(reason="""Mask2Former does not have a get_input_embeddings method""" ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Mask2Former is not a generative model""" ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""Mask2Former does not use token embeddings""" ) def __lowercase ( self : str ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason="""Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def __lowercase ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : int = model_class(A ) UpperCAmelCase__ : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : List[str] = [*signature.parameters.keys()] UpperCAmelCase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,A ) @slow def __lowercase ( self : List[Any] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: UpperCAmelCase__ : Union[str, Any] = MaskaFormerModel.from_pretrained(A ) self.assertIsNotNone(A ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = (self.model_tester.min_size,) * 2 UpperCAmelCase__ : Dict = { """pixel_values""": torch.randn((2, 3, *size) ,device=A ), """mask_labels""": torch.randn((2, 10, *size) ,device=A ), """class_labels""": torch.zeros(2 ,10 ,device=A ).long(), } UpperCAmelCase__ : List[Any] = self.model_tester.get_config() UpperCAmelCase__ : Union[str, Any] = MaskaFormerForUniversalSegmentation(A ).to(A ) UpperCAmelCase__ : List[str] = model(**A ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(A ,**A ,output_hidden_states=A ) def __lowercase ( self : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[int] = model_class(A ).to(A ) UpperCAmelCase__ : str = model(**A ,output_attentions=A ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Optional[int] ): '''simple docstring''' if not self.model_tester.is_training: return UpperCAmelCase__ : Optional[int] = self.all_model_classes[1] UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() UpperCAmelCase__ : Union[str, Any] = model_class(A ) model.to(A ) model.train() UpperCAmelCase__ : Optional[Any] = model(A ,mask_labels=A ,class_labels=A ).loss loss.backward() def __lowercase ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = self.all_model_classes[1] UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : int = model_class(A ).to(A ) model.train() UpperCAmelCase__ : Tuple = model(A ,mask_labels=A ,class_labels=A ) UpperCAmelCase__ : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() UpperCAmelCase__ : Tuple = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() UpperCAmelCase__ : int = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() UpperCAmelCase__ : Union[str, Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=A ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __UpperCAmelCase = 1E-4 def lowerCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class __lowercase ( unittest.TestCase ): @cached_property def __lowercase ( self : str ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : str ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(A ) UpperCAmelCase__ : str = self.default_image_processor UpperCAmelCase__ : List[Any] = prepare_img() UpperCAmelCase__ : Tuple = image_processor(A ,return_tensors="""pt""" ).to(A ) UpperCAmelCase__ : Optional[Any] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(A ,(1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase__ : int = model(**A ) UpperCAmelCase__ : List[Any] = torch.tensor( [[-0.2_7_9_0, -1.0_7_1_7, -1.1_6_6_8], [-0.5_1_2_8, -0.3_1_2_8, -0.4_9_8_7], [-0.5_8_3_2, 0.1_9_7_1, -0.0_1_9_7]] ).to(A ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,A ,atol=A ) ) UpperCAmelCase__ : Union[str, Any] = torch.tensor( [[0.8_9_7_3, 1.1_8_4_7, 1.1_7_7_6], [1.1_9_3_4, 1.5_0_4_0, 1.5_1_2_8], [1.1_1_5_3, 1.4_4_8_6, 1.4_9_5_1]] ).to(A ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,A ,atol=A ) ) UpperCAmelCase__ : int = torch.tensor( [[2.1_1_5_2, 1.7_0_0_0, -0.8_6_0_3], [1.5_8_0_8, 1.8_0_0_4, -0.9_3_5_3], [1.6_0_4_3, 1.7_4_9_5, -0.5_9_9_9]] ).to(A ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,A ,atol=A ) ) def __lowercase ( self : str ): '''simple docstring''' UpperCAmelCase__ : int = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(A ).eval() UpperCAmelCase__ : Union[str, Any] = self.default_image_processor UpperCAmelCase__ : List[str] = prepare_img() UpperCAmelCase__ : List[str] = image_processor(A ,return_tensors="""pt""" ).to(A ) UpperCAmelCase__ : Optional[int] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(A ,(1, 3, 384, 384) ) with torch.no_grad(): UpperCAmelCase__ : Optional[Any] = model(**A ) # masks_queries_logits UpperCAmelCase__ : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) UpperCAmelCase__ : Optional[Any] = [ [-8.7_8_3_9, -9.0_0_5_6, -8.8_1_2_1], [-7.4_1_0_4, -7.0_3_1_3, -6.5_4_0_1], [-6.6_1_0_5, -6.3_4_2_7, -6.4_6_7_5], ] UpperCAmelCase__ : Dict = torch.tensor(A ).to(A ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,A ,atol=A ) ) # class_queries_logits UpperCAmelCase__ : Any = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) UpperCAmelCase__ : str = torch.tensor( [ [1.8_3_2_4, -8.0_8_3_5, -4.1_9_2_2], [0.8_4_5_0, -9.0_0_5_0, -3.6_0_5_3], [0.3_0_4_5, -7.7_2_9_3, -3.0_2_7_5], ] ).to(A ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,A ,atol=A ) ) def __lowercase ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : int = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(A ).eval() UpperCAmelCase__ : str = self.default_image_processor UpperCAmelCase__ : Any = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors="""pt""" ,) UpperCAmelCase__ : Tuple = inputs["""pixel_values"""].to(A ) UpperCAmelCase__ : List[Any] = [el.to(A ) for el in inputs["""mask_labels"""]] UpperCAmelCase__ : Union[str, Any] = [el.to(A ) for el in inputs["""class_labels"""]] with torch.no_grad(): UpperCAmelCase__ : Optional[int] = model(**A ) self.assertTrue(outputs.loss is not None )
65
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : str , __UpperCamelCase : Any ): '''simple docstring''' snake_case_ : Optional[Any] = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] snake_case_ : Optional[int] = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } snake_case_ : Optional[Any] = F'{src_lang}-{tgt_lang}' snake_case_ : Dict = F'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) snake_case_ : List[str] = os.path.join(__UpperCamelCase , """README.md""" ) print(F'Generating {path}' ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(__UpperCamelCase ) # make sure we are under the root of the project __lowerCAmelCase : str = Path(__file__).resolve().parent.parent.parent __lowerCAmelCase : Optional[int] = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[int] = model_name.split('''-''') __lowerCAmelCase : Optional[int] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
58
0
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Optional[int] = abs(SCREAMING_SNAKE_CASE ) _lowercase : Dict = 0 while n > 0: res += n % 10 n //= 10 return res def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Union[str, Any] = abs(SCREAMING_SNAKE_CASE ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: return sum(int(SCREAMING_SNAKE_CASE ) for c in str(abs(SCREAMING_SNAKE_CASE ) ) ) def __magic_name__ ( ) -> None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> None: _lowercase : Tuple = F"""{func.__name__}({value})""" _lowercase : str = timeit(F"""__main__.{call}""" , setup='import __main__' ) print(F"""{call:56} = {func(SCREAMING_SNAKE_CASE )} -- {timing:.4f} seconds""" ) for value in (262_144, 1_125_899_906_842_624, 1_267_650_600_228_229_401_496_703_205_376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
66
"""simple docstring""" __lowerCAmelCase : Tuple = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowerCAmelCase : Tuple = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __lowerCAmelCase : Any = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
58
0
from manim import * class A_ ( UpperCAmelCase ): """simple docstring""" def __UpperCAmelCase ( self : str ) -> Union[str, Any]: _lowercase = Rectangle(height=0.5 ,width=0.5 ) _lowercase = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0 ) _lowercase = Rectangle(height=0.25 ,width=0.25 ) _lowercase = [mem.copy() for i in range(6 )] _lowercase = [mem.copy() for i in range(6 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(__A ,__A ).arrange(__A ,buff=0 ) _lowercase = Text('CPU' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__A ) _lowercase = [mem.copy() for i in range(4 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = Text('GPU' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) gpu.move_to([-1, -1, 0] ) self.add(__A ) _lowercase = [mem.copy() for i in range(6 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = Text('Model' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) model.move_to([3, -1.0, 0] ) self.add(__A ) _lowercase = [] _lowercase = [] for i, rect in enumerate(__A ): _lowercase = fill.copy().set_fill(__A ,opacity=0.8 ) target.move_to(__A ) model_arr.append(__A ) _lowercase = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0.0 ).set_fill(__A ,opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__A ) self.add(*__A ,*__A ) _lowercase = [meta_mem.copy() for i in range(6 )] _lowercase = [meta_mem.copy() for i in range(6 )] _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(*__A ).arrange(__A ,buff=0 ) _lowercase = VGroup(__A ,__A ).arrange(__A ,buff=0 ) _lowercase = Text('Disk' ,font_size=24 ) _lowercase = Group(__A ,__A ).arrange(__A ,buff=0.5 ,aligned_edge=__A ) disk.move_to([-4, -1.25, 0] ) self.add(__A ,__A ) _lowercase = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _lowercase = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" ,font_size=18 ,) key_text.move_to([-5, 2.4, 0] ) self.add(__A ,__A ) _lowercase = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" ,font_size=18 ,) blue_text.next_to(__A ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) self.add(__A ) _lowercase = MarkupText( F"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(__A ) ) _lowercase = Square(0.3 ) input.set_fill(__A ,opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] ,__A ,buff=0.5 ) self.play(Write(__A ) ) input.generate_target() input.target.next_to(model_arr[0] ,direction=__A ,buff=0.02 ) self.play(MoveToTarget(__A ) ) self.play(FadeOut(__A ) ) _lowercase = Arrow(start=__A ,end=__A ,color=__A ,buff=0.5 ) a.next_to(model_arr[0].get_left() ,__A ,buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) _lowercase = MarkupText( F"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(__A ,run_time=3 ) ) _lowercase = {'run_time': 1, 'fade_in': True, 'fade_out': True, 'buff': 0.02} self.play( Write(__A ) ,Circumscribe(model_arr[0] ,color=__A ,**__A ) ,Circumscribe(model_cpu_arr[0] ,color=__A ,**__A ) ,Circumscribe(gpu_rect[0] ,color=__A ,**__A ) ,) self.play(MoveToTarget(model_cpu_arr[0] ) ) _lowercase = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 ,__A ,buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) _lowercase = AnimationGroup( FadeOut(__A ,run_time=0.5 ) ,MoveToTarget(__A ,run_time=0.5 ) ,FadeIn(__A ,run_time=0.5 ) ,lag_ratio=0.2 ) self.play(__A ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: _lowercase = 0.7 self.play( Circumscribe(model_arr[i] ,**__A ) ,Circumscribe(cpu_left_col_base[i] ,**__A ) ,Circumscribe(cpu_left_col_base[i + 1] ,color=__A ,**__A ) ,Circumscribe(gpu_rect[0] ,color=__A ,**__A ) ,Circumscribe(model_arr[i + 1] ,color=__A ,**__A ) ,) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) ,MoveToTarget(model_cpu_arr[i + 1] ) ,) else: self.play( MoveToTarget(model_cpu_arr[i] ,run_time=0.7 ) ,MoveToTarget(model_cpu_arr[i + 1] ,run_time=0.7 ) ,) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() ,RIGHT + 0.02 ,buff=0.2 ) self.play( Circumscribe(model_arr[-1] ,color=__A ,**__A ) ,Circumscribe(cpu_left_col_base[-1] ,color=__A ,**__A ) ,Circumscribe(gpu_rect[0] ,color=__A ,**__A ) ,) self.play(MoveToTarget(model_cpu_arr[i] ) ) _lowercase = a_c _lowercase = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] ,RIGHT + 0.02 ,buff=0.5 ) self.play( FadeOut(__A ) ,FadeOut(__A ,run_time=0.5 ) ,) _lowercase = MarkupText(F"""Inference on a model too large for GPU memory\nis successfully completed.""" ,font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(__A ,run_time=3 ) ,MoveToTarget(__A ) ) self.wait()
67
"""simple docstring""" from jiwer import compute_measures import datasets __lowerCAmelCase : Tuple = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __lowerCAmelCase : Union[str, Any] = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' __lowerCAmelCase : Optional[int] = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", ] , ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=False ) -> Optional[Any]: '''simple docstring''' if concatenate_texts: return compute_measures(_lowercase , _lowercase )["wer"] else: snake_case_ : List[str] = 0 snake_case_ : Optional[int] = 0 for prediction, reference in zip(_lowercase , _lowercase ): snake_case_ : Optional[Any] = compute_measures(_lowercase , _lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
58
0
def lowercase__ ( A_: int = 1 , A_: int = 1000 ) -> int: """simple docstring""" __UpperCAmelCase =1 __UpperCAmelCase =0 for divide_by_number in range(A_ , digit + 1 ): __UpperCAmelCase =[] __UpperCAmelCase =numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(A_ ): __UpperCAmelCase =len(A_ ) __UpperCAmelCase =divide_by_number else: has_been_divided.append(A_ ) __UpperCAmelCase =now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
68
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1_3 , _lowercase=3 , _lowercase=2_2_4 , _lowercase=3_0 , _lowercase=4_0_0 , _lowercase=True , _lowercase=None , _lowercase=True , _lowercase=[0.5, 0.5, 0.5] , _lowercase=[0.5, 0.5, 0.5] , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : Optional[Any] = batch_size snake_case_ : Dict = num_channels snake_case_ : Optional[Any] = image_size snake_case_ : Optional[Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[int] = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : Dict = image_std def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ViTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = EfficientFormerImageProcessorTester(self ) @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.image_proc_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , """image_mean""" ) ) self.assertTrue(hasattr(_lowercase , """image_std""" ) ) self.assertTrue(hasattr(_lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowercase , """do_resize""" ) ) self.assertTrue(hasattr(_lowercase , """size""" ) ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' pass def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input snake_case_ : Any = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Optional[Any] = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Dict = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input snake_case_ : int = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : int = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input snake_case_ : Dict = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Tuple = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
58
0
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float(moles / volume ) * nfactor ) def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float((moles * 0.0821 * temperature) / (volume) ) ) def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
69
"""simple docstring""" from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCAmelCase : int = TypeVar('''KT''') __lowerCAmelCase : Union[str, Any] = TypeVar('''VT''') class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = "root" , _lowercase = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = key snake_case_ : Tuple = value snake_case_ : list[Node[KT, VT]] = [] def __repr__( self ) -> str: '''simple docstring''' return f'Node({self.key}: {self.value})' @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return len(self.forward ) class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = 0.5 , _lowercase = 1_6 ) -> int: '''simple docstring''' snake_case_ : Node[KT, VT] = Node[KT, VT]() snake_case_ : Union[str, Any] = 0 snake_case_ : Optional[int] = p snake_case_ : Any = max_level def __str__( self ) -> str: '''simple docstring''' snake_case_ : str = list(self ) if len(_lowercase ) == 0: return f'SkipList(level={self.level})' snake_case_ : List[Any] = max((len(str(_lowercase ) ) for item in items) , default=4 ) snake_case_ : str = max(_lowercase , 4 ) + 4 snake_case_ : Union[str, Any] = self.head snake_case_ : Dict = [] snake_case_ : List[str] = node.forward.copy() lines.append(f'[{node.key}]'.ljust(_lowercase , """-""" ) + """* """ * len(_lowercase ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) while len(node.forward ) != 0: snake_case_ : Optional[Any] = node.forward[0] lines.append( f'[{node.key}]'.ljust(_lowercase , """-""" ) + """ """.join(str(n.key ) if n.key == node.key else """|""" for n in forwards ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) snake_case_ : List[str] = node.forward lines.append("""None""".ljust(_lowercase ) + """* """ * len(_lowercase ) ) return f'SkipList(level={self.level})\n' + "\n".join(_lowercase ) def __iter__( self ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.head while len(node.forward ) != 0: yield node.forward[0].key snake_case_ : Dict = node.forward[0] def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[int] = 1 while random() < self.p and level < self.max_level: level += 1 return level def UpperCAmelCase__ ( self , _lowercase ) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: snake_case_ : List[Any] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(_lowercase ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: for i, update_node in enumerate(_lowercase ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: snake_case_ : List[str] = node.forward[i] else: snake_case_ : Tuple = update_node.forward[:i] def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> str: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: snake_case_ : List[Any] = value else: snake_case_ : Optional[int] = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , _lowercase ): update_vector.append(self.head ) snake_case_ : Any = level snake_case_ : Optional[int] = Node(_lowercase , _lowercase ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(_lowercase ) else: snake_case_ : Optional[Any] = new_node def UpperCAmelCase__ ( self , _lowercase ) -> VT | None: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: return node.value return None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : List[str] = SkipList() skip_list.insert("""Key1""" , 3 ) skip_list.insert("""Key2""" , 1_2 ) skip_list.insert("""Key3""" , 4_1 ) skip_list.insert("""Key4""" , -1_9 ) snake_case_ : Optional[int] = skip_list.head snake_case_ : List[Any] = {} while node.level != 0: snake_case_ : List[str] = node.forward[0] snake_case_ : Union[str, Any] = node.value assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 1_2 assert all_values["Key3"] == 4_1 assert all_values["Key4"] == -1_9 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_0 ) skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""Key5""" , 7 ) skip_list.insert("""Key7""" , 1_0 ) skip_list.insert("""Key10""" , 5 ) skip_list.insert("""Key7""" , 7 ) skip_list.insert("""Key5""" , 5 ) skip_list.insert("""Key10""" , 1_0 ) snake_case_ : str = skip_list.head snake_case_ : str = {} while node.level != 0: snake_case_ : Optional[Any] = node.forward[0] snake_case_ : int = node.value if len(__UpperCamelCase ) != 4: print() assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 1_2 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 1_0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : str = SkipList() assert skip_list.find("""Some key""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[Any] = SkipList() skip_list.insert("""Key2""" , 2_0 ) assert skip_list.find("""Key2""" ) == 2_0 skip_list.insert("""Some Key""" , 1_0 ) skip_list.insert("""Key2""" , 8 ) skip_list.insert("""V""" , 1_3 ) assert skip_list.find("""Y""" ) is None assert skip_list.find("""Key2""" ) == 8 assert skip_list.find("""Some Key""" ) == 1_0 assert skip_list.find("""V""" ) == 1_3 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Any = SkipList() skip_list.delete("""Some key""" ) assert len(skip_list.head.forward ) == 0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Tuple = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) == 1_4 assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""X""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key1""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Union[str, Any] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4_2 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""X""" ) def traverse_keys(__UpperCamelCase : str ): yield node.key for forward_node in node.forward: yield from traverse_keys(__UpperCamelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __lowerCAmelCase ( ): '''simple docstring''' def is_sorted(__UpperCamelCase : List[Any] ): return all(next_item >= item for item, next_item in zip(__UpperCamelCase , lst[1:] ) ) snake_case_ : str = SkipList() for i in range(1_0 ): skip_list.insert(__UpperCamelCase , __UpperCamelCase ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.insert(-1_2 , -1_2 ) skip_list.insert(7_7 , 7_7 ) assert is_sorted(list(__UpperCamelCase ) ) def __lowerCAmelCase ( ): '''simple docstring''' for _ in range(1_0_0 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = SkipList() skip_list.insert(2 , """2""" ) skip_list.insert(4 , """4""" ) skip_list.insert(6 , """4""" ) skip_list.insert(4 , """5""" ) skip_list.insert(8 , """4""" ) skip_list.insert(9 , """4""" ) skip_list.delete(4 ) print(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
58
0
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase : Any = random.Random() def _SCREAMING_SNAKE_CASE ( lowercase : Union[str, Any] , lowercase : int=1.0 , lowercase : List[str]=None , lowercase : str=None ): '''simple docstring''' if rng is None: lowerCamelCase_ = global_rng lowerCamelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A( unittest.TestCase ): '''simple docstring''' def __init__( self : List[Any] , A_ : Dict , A_ : int=7 , A_ : str=400 , A_ : Dict=2000 , A_ : List[Any]=24 , A_ : List[Any]=24 , A_ : int=0.0 , A_ : Dict=16000 , A_ : List[Any]=True , A_ : str=True , ) -> Dict: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = min_seq_length lowerCamelCase_ = max_seq_length lowerCamelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase_ = feature_size lowerCamelCase_ = num_mel_bins lowerCamelCase_ = padding_value lowerCamelCase_ = sampling_rate lowerCamelCase_ = return_attention_mask lowerCamelCase_ = do_normalize def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a__ ( self : List[Any] , A_ : str=False , A_ : Union[str, Any]=False ) -> str: """simple docstring""" def _flatten(A_ : List[Any] ): return list(itertools.chain(*A_ ) ) if equal_length: lowerCamelCase_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCamelCase_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A( UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = SpeechaTextFeatureExtractionTester(self ) def a__ ( self : str , A_ : Dict ) -> Dict: """simple docstring""" self.assertTrue(np.all(np.mean(A_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(A_ , axis=0 ) - 1 ) < 1E-3 ) ) def a__ ( self : int ) -> List[Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test feature size lowerCamelCase_ = feature_extractor(A_ , padding=A_ , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCamelCase_ = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test batched lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase_ = np.asarray(A_ ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features lowerCamelCase_ = feature_extractor(A_ , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(A_ , A_ ): self.assertTrue(np.allclose(A_ , A_ , atol=1E-3 ) ) def a__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , padding=A_ , max_length=A_ , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[Any] ) -> Any: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase_ = [None, 16, None] for max_length, padding in zip(A_ , A_ ): lowerCamelCase_ = feature_extractor( A_ , max_length=A_ , padding=A_ , return_tensors='np' , return_attention_mask=A_ ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = [np.sum(A_ ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def a__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='max_length' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def a__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=4 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCamelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase_ = feature_extractor( A_ , padding='longest' , max_length=16 , truncation=A_ , return_tensors='np' , return_attention_mask=A_ , ) lowerCamelCase_ = inputs.input_features lowerCamelCase_ = inputs.attention_mask lowerCamelCase_ = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def a__ ( self : List[Any] ) -> List[str]: """simple docstring""" import torch lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCamelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCamelCase_ = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def a__ ( self : List[str] , A_ : Union[str, Any] ) -> List[Any]: """simple docstring""" from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech lowerCamelCase_ = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def a__ ( self : str ) -> Tuple: """simple docstring""" lowerCamelCase_ = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCamelCase_ = self._load_datasamples(1 ) lowerCamelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase_ = feature_extractor(A_ , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , A_ , atol=1E-4 ) )
70
"""simple docstring""" import argparse import os import re import packaging.version __lowerCAmelCase : Optional[Any] = '''examples/''' __lowerCAmelCase : Union[str, Any] = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __lowerCAmelCase : Union[str, Any] = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } __lowerCAmelCase : List[Any] = '''README.md''' def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : Dict ): '''simple docstring''' with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : Any = f.read() snake_case_ , snake_case_ : Optional[int] = REPLACE_PATTERNS[pattern] snake_case_ : Union[str, Any] = replace.replace("""VERSION""" , __UpperCamelCase ) snake_case_ : List[Any] = re_pattern.sub(__UpperCamelCase , __UpperCamelCase ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' for folder, directories, fnames in os.walk(__UpperCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(__UpperCamelCase , __UpperCamelCase ) , __UpperCamelCase , pattern="""examples""" ) def __lowerCAmelCase ( __UpperCamelCase : List[str] , __UpperCamelCase : int=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not patch: update_version_in_examples(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = """🤗 Transformers currently provides the following architectures""" snake_case_ : Union[str, Any] = """1. Want to contribute a new model?""" with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : str = f.readlines() # Find the start of the list. snake_case_ : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 snake_case_ : Optional[int] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): snake_case_ : Any = lines[index].replace( """https://huggingface.co/docs/diffusers/main/model_doc""" , """https://huggingface.co/docs/diffusers/model_doc""" , ) index += 1 with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] , """r""" ) as f: snake_case_ : Any = f.read() snake_case_ : Tuple = REPLACE_PATTERNS["""init"""][0].search(__UpperCamelCase ).groups()[0] return packaging.version.parse(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : str=False ): '''simple docstring''' snake_case_ : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: snake_case_ : str = default_version.base_version elif patch: snake_case_ : str = F'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: snake_case_ : str = F'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. snake_case_ : int = input(F'Which version are you releasing? [{default_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Optional[int] = default_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase , patch=__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = get_version() snake_case_ : str = F'{current_version.major}.{current_version.minor + 1}.0.dev0' snake_case_ : Tuple = current_version.base_version # Check with the user we got that right. snake_case_ : Optional[int] = input(F'Which version are we developing now? [{dev_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Dict = dev_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": __lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __lowerCAmelCase : str = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
58
0
'''simple docstring''' import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging _lowerCamelCase = logging.get_logger(__name__) def a__ ( _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : List[str]=False ) -> List[str]: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( "Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see" " https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation" " instructions." ) raise if not is_sharded: UpperCAmelCase_ : Dict = os.path.abspath(_SCREAMING_SNAKE_CASE ) logger.info(F'''Loading PyTorch weights from {pt_path}''' ) UpperCAmelCase_ : List[str] = torch.load(_SCREAMING_SNAKE_CASE , map_location="cpu" ) logger.info(F'''PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.''' ) UpperCAmelCase_ : str = convert_pytorch_state_dict_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files UpperCAmelCase_ : List[str] = convert_pytorch_sharded_state_dict_to_flax(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return flax_state_dict def a__ ( _SCREAMING_SNAKE_CASE : Tuple[str] , _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : Dict[str, jnp.ndarray] , _SCREAMING_SNAKE_CASE : str , ) -> (Tuple[str], np.ndarray): """simple docstring""" def is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE : Tuple[str] ) -> bool: return len(set(_SCREAMING_SNAKE_CASE ) & {key, (model_prefix,) + key} ) > 0 # layer norm UpperCAmelCase_ : List[Any] = pt_tuple_key[:-1] + ("scale",) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean UpperCAmelCase_ : Dict = pt_tuple_key[:-1] + ("mean",) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var UpperCAmelCase_ : Any = pt_tuple_key[:-1] + ("var",) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE ): return renamed_pt_tuple_key, pt_tensor # embedding UpperCAmelCase_ : str = pt_tuple_key[:-1] + ("embedding",) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE ): return renamed_pt_tuple_key, pt_tensor # conv layer UpperCAmelCase_ : Tuple = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : str = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer UpperCAmelCase_ : Optional[int] = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(_SCREAMING_SNAKE_CASE ): UpperCAmelCase_ : Optional[Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight UpperCAmelCase_ : Optional[Any] = pt_tuple_key[:-1] + ("weight",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias UpperCAmelCase_ : Optional[Any] = pt_tuple_key[:-1] + ("bias",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 UpperCAmelCase_ : Optional[int] = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): UpperCAmelCase_ : Union[str, Any] = pt_tuple_key[-2] + "_g" elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): UpperCAmelCase_ : str = pt_tuple_key[-2] + "_v" if name is not None: UpperCAmelCase_ : int = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def a__ ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : Any ) -> str: """simple docstring""" UpperCAmelCase_ : Tuple = {k: v.numpy() for k, v in pt_state_dict.items()} UpperCAmelCase_ : List[str] = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: UpperCAmelCase_ : List[Any] = flax_model.params["params"] else: UpperCAmelCase_ : Optional[Any] = flax_model.params UpperCAmelCase_ : str = flatten_dict(_SCREAMING_SNAKE_CASE ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: UpperCAmelCase_ : Tuple = flatten_dict(flax_model.params["batch_stats"] ) random_flax_state_dict.update(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = {} UpperCAmelCase_ : List[str] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split("." )[0] for k in pt_state_dict.keys()} ) UpperCAmelCase_ : str = (model_prefix in flax_model_params) and ( model_prefix not in {k.split("." )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): UpperCAmelCase_ : Dict = tuple(pt_key.split("." ) ) # remove base model prefix if necessary UpperCAmelCase_ : List[Any] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: UpperCAmelCase_ : int = pt_tuple_key[1:] # Correctly rename weight parameters UpperCAmelCase_ , UpperCAmelCase_ : Tuple = rename_key_and_reshape_tensor( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # add model prefix if necessary UpperCAmelCase_ : Any = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: UpperCAmelCase_ : Tuple = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'''PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ''' F'''{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: UpperCAmelCase_ : str = jnp.asarray(_SCREAMING_SNAKE_CASE ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue # also add unexpected weight so that warning is thrown UpperCAmelCase_ : List[Any] = jnp.asarray(_SCREAMING_SNAKE_CASE ) else: # also add unexpected weight so that warning is thrown UpperCAmelCase_ : Optional[int] = jnp.asarray(_SCREAMING_SNAKE_CASE ) return unflatten_dict(_SCREAMING_SNAKE_CASE ) def a__ ( _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : str ) -> Optional[Any]: """simple docstring""" import torch # Load the index UpperCAmelCase_ : Optional[Any] = {} for shard_file in shard_filenames: # load using msgpack utils UpperCAmelCase_ : str = torch.load(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = {k: v.numpy() for k, v in pt_state_dict.items()} UpperCAmelCase_ : Optional[int] = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: UpperCAmelCase_ : List[str] = flax_model.params["params"] UpperCAmelCase_ : List[Any] = flatten_dict(_SCREAMING_SNAKE_CASE ) random_flax_state_dict.update(flatten_dict(flax_model.params["batch_stats"] ) ) else: UpperCAmelCase_ : Tuple = flax_model.params UpperCAmelCase_ : Any = flatten_dict(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split("." )[0] for k in pt_state_dict.keys()} ) UpperCAmelCase_ : List[Any] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split("." )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): UpperCAmelCase_ : Optional[int] = tuple(pt_key.split("." ) ) # remove base model prefix if necessary UpperCAmelCase_ : Optional[int] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: UpperCAmelCase_ : List[Any] = pt_tuple_key[1:] # Correctly rename weight parameters UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = rename_key_and_reshape_tensor( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # add model prefix if necessary UpperCAmelCase_ : Optional[Any] = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: UpperCAmelCase_ : Optional[Any] = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'''PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ''' F'''{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: UpperCAmelCase_ : Dict = jnp.asarray(_SCREAMING_SNAKE_CASE ) continue if "var" in flax_key[-1]: UpperCAmelCase_ : Dict = jnp.asarray(_SCREAMING_SNAKE_CASE ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue # also add unexpected weight so that warning is thrown UpperCAmelCase_ : str = jnp.asarray(_SCREAMING_SNAKE_CASE ) else: # also add unexpected weight so that warning is thrown UpperCAmelCase_ : Optional[int] = jnp.asarray(_SCREAMING_SNAKE_CASE ) return unflatten_dict(_SCREAMING_SNAKE_CASE ) def a__ ( _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase_ : List[str] = os.path.abspath(_SCREAMING_SNAKE_CASE ) logger.info(F'''Loading Flax weights from {flax_checkpoint_path}''' ) # import correct flax class UpperCAmelCase_ : Optional[int] = getattr(_SCREAMING_SNAKE_CASE , "Flax" + model.__class__.__name__ ) # load flax weight dict with open(_SCREAMING_SNAKE_CASE , "rb" ) as state_f: try: UpperCAmelCase_ : Union[str, Any] = from_bytes(_SCREAMING_SNAKE_CASE , state_f.read() ) except UnpicklingError: raise EnvironmentError(F'''Unable to convert {flax_checkpoint_path} to Flax deserializable object. ''' ) return load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def a__ ( _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Dict ) -> str: """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( "Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see" " https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation" " instructions." ) raise # check if we have bf16 weights UpperCAmelCase_ : List[str] = flatten_dict(jax.tree_util.tree_map(lambda _SCREAMING_SNAKE_CASE : x.dtype == jnp.bfloataa , _SCREAMING_SNAKE_CASE ) ).values() if any(_SCREAMING_SNAKE_CASE ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( "Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` " "before loading those in PyTorch model." ) UpperCAmelCase_ : List[Any] = jax.tree_util.tree_map( lambda _SCREAMING_SNAKE_CASE : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , _SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = flatten_dict(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = pt_model.state_dict() UpperCAmelCase_ : Optional[Any] = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split("." )[0] for k in pt_model_dict.keys()} ) UpperCAmelCase_ : Any = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split("." )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys UpperCAmelCase_ : Optional[Any] = [] UpperCAmelCase_ : Optional[int] = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): UpperCAmelCase_ : List[str] = flax_key_tuple[0] == pt_model.base_model_prefix UpperCAmelCase_ : Dict = ".".join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: UpperCAmelCase_ : Union[str, Any] = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: UpperCAmelCase_ : Any = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(_SCREAMING_SNAKE_CASE ) not in pt_model_dict: # conv layer UpperCAmelCase_ : Dict = flax_key_tuple[:-1] + ("weight",) UpperCAmelCase_ : Optional[Any] = jnp.transpose(_SCREAMING_SNAKE_CASE , (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(_SCREAMING_SNAKE_CASE ) not in pt_model_dict: # linear layer UpperCAmelCase_ : List[str] = flax_key_tuple[:-1] + ("weight",) UpperCAmelCase_ : Tuple = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: UpperCAmelCase_ : Optional[int] = flax_key_tuple[:-1] + ("weight",) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: UpperCAmelCase_ : Dict = flax_key_tuple[:-1] + ("running_mean",) elif "var" in flax_key_tuple[-1]: UpperCAmelCase_ : Tuple = flax_key_tuple[:-1] + ("running_var",) if "batch_stats" in flax_state: UpperCAmelCase_ : Dict = ".".join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: UpperCAmelCase_ : Union[str, Any] = ".".join(_SCREAMING_SNAKE_CASE ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. UpperCAmelCase_ : Dict = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: UpperCAmelCase_ : List[Any] = key.split("." ) UpperCAmelCase_ : int = None if key_components[-3::2] == ["parametrizations", "original0"]: UpperCAmelCase_ : str = key_components[-2] + "_g" elif key_components[-3::2] == ["parametrizations", "original1"]: UpperCAmelCase_ : Dict = key_components[-2] + "_v" if name is not None: UpperCAmelCase_ : Dict = key_components[:-3] + [name] UpperCAmelCase_ : Union[str, Any] = ".".join(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Any = key if flax_key in special_pt_names: UpperCAmelCase_ : List[str] = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'''Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ''' F'''to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) else: # add weight to pytorch dict UpperCAmelCase_ : List[str] = np.asarray(_SCREAMING_SNAKE_CASE ) if not isinstance(_SCREAMING_SNAKE_CASE , np.ndarray ) else flax_tensor UpperCAmelCase_ : List[Any] = torch.from_numpy(_SCREAMING_SNAKE_CASE ) # remove from missing keys missing_keys.remove(_SCREAMING_SNAKE_CASE ) else: # weight is not expected by PyTorch model unexpected_keys.append(_SCREAMING_SNAKE_CASE ) pt_model.load_state_dict(_SCREAMING_SNAKE_CASE ) # re-transform missing_keys to list UpperCAmelCase_ : Tuple = list(_SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: logger.warning( "Some weights of the Flax model were not used when initializing the PyTorch model" F''' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing''' F''' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture''' " (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This" F''' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect''' " to be exactly identical (e.g. initializing a BertForSequenceClassification model from a" " FlaxBertForSequenceClassification model)." ) else: logger.warning(F'''All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n''' ) if len(_SCREAMING_SNAKE_CASE ) > 0: logger.warning( F'''Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly''' F''' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to''' " use it for predictions and inference." ) else: logger.warning( F'''All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n''' "If your task is similar to the task the model of the checkpoint was trained on, " F'''you can already use {pt_model.__class__.__name__} for predictions without further training.''' ) return pt_model
71
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
58
0
'''simple docstring''' class __magic_name__ : def __init__( self , snake_case_ , snake_case_ ): lowercase =name lowercase =val def __str__( self ): return f'{self.__class__.__name__}({self.name}, {self.val})' def __lt__( self , snake_case_ ): return self.val < other.val class __magic_name__ : def __init__( self , snake_case_ ): lowercase ={} lowercase ={} lowercase =self.build_heap(snake_case_ ) def __getitem__( self , snake_case_ ): return self.get_value(snake_case_ ) def _A( self , snake_case_ ): return (idx - 1) // 2 def _A( self , snake_case_ ): return idx * 2 + 1 def _A( self , snake_case_ ): return idx * 2 + 2 def _A( self , snake_case_ ): return self.heap_dict[key] def _A( self , snake_case_ ): lowercase =len(snake_case_ ) - 1 lowercase =self.get_parent_idx(snake_case_ ) for idx, i in enumerate(snake_case_ ): lowercase =idx lowercase =i.val for i in range(snake_case_ , -1 , -1 ): self.sift_down(snake_case_ , snake_case_ ) return array def _A( self , snake_case_ , snake_case_ ): while True: lowercase =self.get_left_child_idx(snake_case_ ) # noqa: E741 lowercase =self.get_right_child_idx(snake_case_ ) lowercase =idx if l < len(snake_case_ ) and array[l] < array[idx]: lowercase =l if r < len(snake_case_ ) and array[r] < array[smallest]: lowercase =r if smallest != idx: lowercase , lowercase =array[smallest], array[idx] ( ( lowercase ) , ( lowercase ) , ) =( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) lowercase =smallest else: break def _A( self , snake_case_ ): lowercase =self.get_parent_idx(snake_case_ ) while p >= 0 and self.heap[p] > self.heap[idx]: lowercase , lowercase =self.heap[idx], self.heap[p] lowercase , lowercase =( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) lowercase =p lowercase =self.get_parent_idx(snake_case_ ) def _A( self ): return self.heap[0] def _A( self ): lowercase , lowercase =self.heap[-1], self.heap[0] lowercase , lowercase =( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) lowercase =self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def _A( self , snake_case_ ): self.heap.append(snake_case_ ) lowercase =len(self.heap ) - 1 lowercase =node.val self.sift_up(len(self.heap ) - 1 ) def _A( self ): return len(self.heap ) == 0 def _A( self , snake_case_ , snake_case_ ): assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" lowercase =new_value lowercase =new_value self.sift_up(self.idx_of_element[node] ) _UpperCAmelCase : Any = Node('''R''', -1) _UpperCAmelCase : Optional[int] = Node('''B''', 6) _UpperCAmelCase : Tuple = Node('''A''', 3) _UpperCAmelCase : Union[str, Any] = Node('''X''', 1) _UpperCAmelCase : List[str] = Node('''E''', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array _UpperCAmelCase : Union[str, Any] = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('''Min Heap - before decrease key''') for i in my_min_heap.heap: print(i) print('''Min Heap - After decrease key of node [B -> -17]''') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
72
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) snake_case_ : str = precision snake_case_ : Any = ceil(precision / 1_4 ) snake_case_ : Dict = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : Optional[Any] = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : Optional[int] = Decimal(__UpperCamelCase ) for k in range(1 , __UpperCamelCase ): snake_case_ : Any = factorial(6 * k ) // (factorial(3 * k ) * factorial(__UpperCamelCase ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __lowerCAmelCase : int = 50 print(F'''The first {n} digits of pi is: {pi(n)}''')
58
0
import numpy as np def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 1e-12 , _UpperCAmelCase = 100 , ): assert np.shape(_UpperCAmelCase)[0] == np.shape(_UpperCAmelCase)[1] # Ensure proper dimensionality. assert np.shape(_UpperCAmelCase)[0] == np.shape(_UpperCAmelCase)[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(_UpperCAmelCase) == np.iscomplexobj(_UpperCAmelCase) SCREAMING_SNAKE_CASE = np.iscomplexobj(_UpperCAmelCase) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(_UpperCAmelCase , input_matrix.conj().T) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 1e12 while not convergence: # Multiple matrix by the vector. SCREAMING_SNAKE_CASE = np.dot(_UpperCAmelCase , _UpperCAmelCase) # Normalize the resulting output vector. SCREAMING_SNAKE_CASE = w / np.linalg.norm(_UpperCAmelCase) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) SCREAMING_SNAKE_CASE = vector.conj().T if is_complex else vector.T SCREAMING_SNAKE_CASE = np.dot(_UpperCAmelCase , np.dot(_UpperCAmelCase , _UpperCAmelCase)) # Check convergence. SCREAMING_SNAKE_CASE = np.abs(lambda_ - lambda_previous) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = lambda_ if is_complex: SCREAMING_SNAKE_CASE = np.real(lambda_) return lambda_, vector def lowerCamelCase__ (): SCREAMING_SNAKE_CASE = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]]) SCREAMING_SNAKE_CASE = np.array([41, 4, 20]) SCREAMING_SNAKE_CASE = real_input_matrix.astype(np.complexaaa) SCREAMING_SNAKE_CASE = np.triu(1j * complex_input_matrix , 1) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T SCREAMING_SNAKE_CASE = np.array([41, 4, 20]).astype(np.complexaaa) for problem_type in ["real", "complex"]: if problem_type == "real": SCREAMING_SNAKE_CASE = real_input_matrix SCREAMING_SNAKE_CASE = real_vector elif problem_type == "complex": SCREAMING_SNAKE_CASE = complex_input_matrix SCREAMING_SNAKE_CASE = complex_vector # Our implementation. SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = power_iteration(_UpperCAmelCase , _UpperCAmelCase) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = np.linalg.eigh(_UpperCAmelCase) # Last eigenvalue is the maximum one. SCREAMING_SNAKE_CASE = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. SCREAMING_SNAKE_CASE = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(_UpperCAmelCase) - np.abs(_UpperCAmelCase)) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
73
"""simple docstring""" import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCAmelCase ( __UpperCamelCase : Optional[int] ): '''simple docstring''' snake_case_ : Any = torch.exp(__UpperCamelCase ) snake_case_ : Optional[int] = torch.sum(__UpperCamelCase , dim=1 ) # sum of exp(x_i) snake_case_ : str = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(__UpperCamelCase ) - B / A class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> int: '''simple docstring''' super().__init__() snake_case_ : Tuple = config.output_attentions snake_case_ : str = config.output_hidden_states snake_case_ : List[str] = nn.ModuleList([BertLayer(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Tuple = nn.ModuleList([BertHighway(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Any = [-1 for _ in range(config.num_hidden_layers )] def UpperCAmelCase__ ( self , _lowercase ) -> Tuple: '''simple docstring''' if (type(_lowercase ) is float) or (type(_lowercase ) is int): for i in range(len(self.early_exit_entropy ) ): snake_case_ : Dict = x else: snake_case_ : Union[str, Any] = x def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : int = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Any: '''simple docstring''' snake_case_ : str = () snake_case_ : str = () snake_case_ : List[str] = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: snake_case_ : int = all_hidden_states + (hidden_states,) snake_case_ : Any = layer_module( _lowercase , _lowercase , head_mask[i] , _lowercase , _lowercase ) snake_case_ : Dict = layer_outputs[0] if self.output_attentions: snake_case_ : str = all_attentions + (layer_outputs[1],) snake_case_ : Optional[int] = (hidden_states,) if self.output_hidden_states: snake_case_ : Tuple = current_outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : int = current_outputs + (all_attentions,) snake_case_ : Optional[Any] = self.highway[i](_lowercase ) # logits, pooled_output if not self.training: snake_case_ : Tuple = highway_exit[0] snake_case_ : List[str] = entropy(_lowercase ) snake_case_ : str = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy snake_case_ : Union[str, Any] = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: snake_case_ : List[Any] = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(_lowercase , i + 1 ) else: snake_case_ : Dict = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: snake_case_ : Dict = all_hidden_states + (hidden_states,) snake_case_ : str = (hidden_states,) if self.output_hidden_states: snake_case_ : List[Any] = outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : Union[str, Any] = outputs + (all_attentions,) snake_case_ : List[str] = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( '''The Bert Model transformer with early exiting (DeeBERT). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> Union[str, Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config snake_case_ : int = BertEmbeddings(_lowercase ) snake_case_ : Tuple = DeeBertEncoder(_lowercase ) snake_case_ : int = BertPooler(_lowercase ) self.init_weights() def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' self.encoder.init_highway_pooler(self.pooler ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return self.embeddings.word_embeddings def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = value def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(_lowercase ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' if input_ids is not None and inputs_embeds is not None: raise ValueError("""You cannot specify both input_ids and inputs_embeds at the same time""" ) elif input_ids is not None: snake_case_ : Dict = input_ids.size() elif inputs_embeds is not None: snake_case_ : Optional[int] = inputs_embeds.size()[:-1] else: raise ValueError("""You have to specify either input_ids or inputs_embeds""" ) snake_case_ : int = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: snake_case_ : Dict = torch.ones(_lowercase , device=_lowercase ) if encoder_attention_mask is None: snake_case_ : Tuple = torch.ones(_lowercase , device=_lowercase ) if token_type_ids is None: snake_case_ : Any = torch.zeros(_lowercase , dtype=torch.long , device=_lowercase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. snake_case_ : torch.Tensor = self.get_extended_attention_mask(_lowercase , _lowercase , _lowercase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: snake_case_ : List[str] = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: snake_case_ : Any = encoder_attention_mask[:, None, None, :] snake_case_ : List[str] = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility snake_case_ : List[str] = (1.0 - encoder_extended_attention_mask) * -1_0000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] snake_case_ : int = self.get_head_mask(_lowercase , self.config.num_hidden_layers ) snake_case_ : List[str] = self.embeddings( input_ids=_lowercase , position_ids=_lowercase , token_type_ids=_lowercase , inputs_embeds=_lowercase ) snake_case_ : List[str] = self.encoder( _lowercase , attention_mask=_lowercase , head_mask=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , ) snake_case_ : Optional[Any] = encoder_outputs[0] snake_case_ : Union[str, Any] = self.pooler(_lowercase ) snake_case_ : Optional[Any] = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = message snake_case_ : str = exit_layer # start from 1! class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__() snake_case_ : str = BertPooler(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Dict = nn.Linear(config.hidden_size , config.num_labels ) def UpperCAmelCase__ ( self , _lowercase ) -> Dict: '''simple docstring''' snake_case_ : List[str] = encoder_outputs[0] snake_case_ : List[Any] = self.pooler(_lowercase ) # "return" pooler_output # BertModel snake_case_ : Tuple = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification snake_case_ : Union[str, Any] = bmodel_output[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : List[str] = self.classifier(_lowercase ) return logits, pooled_output @add_start_docstrings( '''Bert Model (with early exiting - DeeBERT) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config.num_labels snake_case_ : Tuple = config.num_hidden_layers snake_case_ : Any = DeeBertModel(_lowercase ) snake_case_ : Optional[int] = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Tuple = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> int: '''simple docstring''' snake_case_ : int = self.num_layers try: snake_case_ : Any = self.bert( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits snake_case_ : str = outputs[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : List[Any] = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : Optional[int] = e.message snake_case_ : Dict = e.exit_layer snake_case_ : Optional[Any] = outputs[0] if not self.training: snake_case_ : int = entropy(_lowercase ) snake_case_ : int = [] snake_case_ : List[str] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : Dict = [] for highway_exit in outputs[-1]: snake_case_ : List[Any] = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : List[Any] = MSELoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : List[str] = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : str = (loss,) + outputs if not self.training: snake_case_ : Optional[int] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : str = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
58
0
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] __SCREAMING_SNAKE_CASE : Any = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } __SCREAMING_SNAKE_CASE : Dict = F'''{src_lang}-{tgt_lang}''' __SCREAMING_SNAKE_CASE : Tuple = F''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(snake_case , exist_ok=snake_case ) __SCREAMING_SNAKE_CASE : List[str] = os.path.join(snake_case , '''README.md''' ) print(F'''Generating {path}''' ) with open(snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write(snake_case ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowercase_ , lowercase_ , lowercase_ = model_name.split("""-""") lowercase_ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
74
"""simple docstring""" # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' return 1 / (1 + np.exp(-z )) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int ): '''simple docstring''' return (-y * np.log(__UpperCamelCase ) - (1 - y) * np.log(1 - h )).mean() def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : List[Any] , __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Optional[int] = np.dot(__UpperCamelCase , __UpperCamelCase ) return np.sum(y * scores - np.log(1 + np.exp(__UpperCamelCase ) ) ) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : int=7_0_0_0_0 ): '''simple docstring''' snake_case_ : Dict = np.zeros(x.shape[1] ) for iterations in range(__UpperCamelCase ): snake_case_ : Any = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Optional[Any] = np.dot(x.T , h - y ) / y.size snake_case_ : str = theta - alpha * gradient # updating the weights snake_case_ : int = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Dict = cost_function(__UpperCamelCase , __UpperCamelCase ) if iterations % 1_0_0 == 0: print(F'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": __lowerCAmelCase : Any = datasets.load_iris() __lowerCAmelCase : List[Any] = iris.data[:, :2] __lowerCAmelCase : Tuple = (iris.target != 0) * 1 __lowerCAmelCase : Any = 0.1 __lowerCAmelCase : List[Any] = logistic_reg(alpha, x, y, max_iterations=7_0000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' return sigmoid_function( np.dot(__UpperCamelCase , __UpperCamelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = (x[:, 0].min(), x[:, 0].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Tuple = (x[:, 1].min(), x[:, 1].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Optional[Any] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) __lowerCAmelCase : Any = np.c_[xxa.ravel(), xxa.ravel()] __lowerCAmelCase : Optional[int] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
58
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor UpperCamelCase__ = logging.get_logger(__name__) class lowerCamelCase_ ( __a ): def __init__( self : int , *_A : Union[str, Any] , **_A : Any ): '''simple docstring''' warnings.warn( '''The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use SegformerImageProcessor instead.''' , _A , ) super().__init__(*_A , **_A )
75
"""simple docstring""" import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput __lowerCAmelCase : Tuple = '''scheduler_config.json''' class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 1 _lowerCamelCase = 2 _lowerCamelCase = 3 _lowerCamelCase = 4 _lowerCamelCase = 5 @dataclass class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 42 class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = SCHEDULER_CONFIG_NAME _lowerCamelCase = ['''dtype'''] _lowerCamelCase = [] _lowerCamelCase = True @classmethod def UpperCAmelCase__ ( cls , _lowercase = None , _lowercase = None , _lowercase=False , **_lowercase , ) -> Any: '''simple docstring''' snake_case_ , snake_case_ : int = cls.load_config( pretrained_model_name_or_path=_lowercase , subfolder=_lowercase , return_unused_kwargs=_lowercase , **_lowercase , ) snake_case_ , snake_case_ : Dict = cls.from_config(_lowercase , return_unused_kwargs=_lowercase , **_lowercase ) if hasattr(_lowercase , """create_state""" ) and getattr(_lowercase , """has_state""" , _lowercase ): snake_case_ : Any = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCAmelCase__ ( self , _lowercase , _lowercase = False , **_lowercase ) -> Optional[Any]: '''simple docstring''' self.save_config(save_directory=_lowercase , push_to_hub=_lowercase , **_lowercase ) @property def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' return self._get_compatibles() @classmethod def UpperCAmelCase__ ( cls ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = list(set([cls.__name__] + cls._compatibles ) ) snake_case_ : str = importlib.import_module(__name__.split(""".""" )[0] ) snake_case_ : Optional[int] = [ getattr(_lowercase , _lowercase ) for c in compatible_classes_str if hasattr(_lowercase , _lowercase ) ] return compatible_classes def __lowerCAmelCase ( __UpperCamelCase : jnp.ndarray , __UpperCamelCase : Tuple[int] ): '''simple docstring''' assert len(__UpperCamelCase ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__UpperCamelCase ) - x.ndim) ) , __UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Any=0.999 , __UpperCamelCase : Optional[int]=jnp.floataa ): '''simple docstring''' def alpha_bar(__UpperCamelCase : Optional[int] ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 snake_case_ : Optional[Any] = [] for i in range(__UpperCamelCase ): snake_case_ : Dict = i / num_diffusion_timesteps snake_case_ : Union[str, Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__UpperCamelCase ) / alpha_bar(__UpperCamelCase ) , __UpperCamelCase ) ) return jnp.array(__UpperCamelCase , dtype=__UpperCamelCase ) @flax.struct.dataclass class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 @classmethod def UpperCAmelCase__ ( cls , _lowercase ) -> int: '''simple docstring''' snake_case_ : Any = scheduler.config if config.trained_betas is not None: snake_case_ : Optional[Any] = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": snake_case_ : int = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case_ : str = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case_ : int = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( f'beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}' ) snake_case_ : Optional[Any] = 1.0 - betas snake_case_ : Any = jnp.cumprod(_lowercase , axis=0 ) return cls( alphas=_lowercase , betas=_lowercase , alphas_cumprod=_lowercase , ) def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ : Tuple = state.alphas_cumprod snake_case_ : Optional[int] = alphas_cumprod[timesteps] ** 0.5 snake_case_ : Dict = sqrt_alpha_prod.flatten() snake_case_ : int = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) snake_case_ : Optional[Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 snake_case_ : Dict = sqrt_one_minus_alpha_prod.flatten() snake_case_ : Tuple = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : str = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : List[Any] = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
58
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'uw-madison/mra-base-512-4': 'https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json', } class UpperCAmelCase_ ( snake_case ): UpperCamelCase ="mra" def __init__( self , UpperCamelCase_=5_02_65 , UpperCamelCase_=7_68 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=30_72 , UpperCamelCase_="gelu" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=5_12 , UpperCamelCase_=1 , UpperCamelCase_=0.0_2 , UpperCamelCase_=1E-5 , UpperCamelCase_="absolute" , UpperCamelCase_=4 , UpperCamelCase_="full" , UpperCamelCase_=0 , UpperCamelCase_=0 , UpperCamelCase_=1 , UpperCamelCase_=0 , UpperCamelCase_=2 , **UpperCamelCase_ , ) -> Optional[Any]: super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_ ) __lowercase : Optional[Any] = vocab_size __lowercase : List[Any] = max_position_embeddings __lowercase : Union[str, Any] = hidden_size __lowercase : Optional[int] = num_hidden_layers __lowercase : Optional[int] = num_attention_heads __lowercase : Union[str, Any] = intermediate_size __lowercase : List[str] = hidden_act __lowercase : Union[str, Any] = hidden_dropout_prob __lowercase : Any = attention_probs_dropout_prob __lowercase : Optional[int] = initializer_range __lowercase : Union[str, Any] = type_vocab_size __lowercase : Any = layer_norm_eps __lowercase : List[str] = position_embedding_type __lowercase : int = block_per_row __lowercase : Union[str, Any] = approx_mode __lowercase : Optional[int] = initial_prior_first_n_blocks __lowercase : str = initial_prior_diagonal_n_blocks
76
"""simple docstring""" from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( '''The RoBERTa Model transformer with early exiting (DeeRoBERTa). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : str = RobertaEmbeddings(_lowercase ) self.init_weights() @add_start_docstrings( '''RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Optional[Any] = config.num_labels snake_case_ : Dict = config.num_hidden_layers snake_case_ : str = DeeRobertaModel(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : List[str] = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> Tuple: '''simple docstring''' snake_case_ : Any = self.num_layers try: snake_case_ : int = self.roberta( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) snake_case_ : str = outputs[1] snake_case_ : Union[str, Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : Dict = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : List[Any] = e.message snake_case_ : Union[str, Any] = e.exit_layer snake_case_ : Dict = outputs[0] if not self.training: snake_case_ : Dict = entropy(_lowercase ) snake_case_ : Optional[int] = [] snake_case_ : Union[str, Any] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Dict = MSELoss() snake_case_ : Dict = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Union[str, Any] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : int = [] for highway_exit in outputs[-1]: snake_case_ : Tuple = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Optional[Any] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Optional[int] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : Dict = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : List[str] = (loss,) + outputs if not self.training: snake_case_ : Optional[Any] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : Tuple = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
58
0
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A = {"""configuration_mmbt""": ["""MMBTConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""MMBTForClassification""", """MMBTModel""", """ModalEmbeddings"""] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : list[int] ): '''simple docstring''' if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : list[int] , __UpperCamelCase : int ): '''simple docstring''' if curr_ind == len(__UpperCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(__UpperCamelCase ) ): if valid_connection(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): # Insert current vertex into path as next transition snake_case_ : List[str] = next_ver # Validate created path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , curr_ind + 1 ): return True # Backtrack snake_case_ : Tuple = -1 return False def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int = 0 ): '''simple docstring''' snake_case_ : Tuple = [-1] * (len(__UpperCamelCase ) + 1) # initialize start and end of path with starting index snake_case_ : Optional[int] = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , 1 ) else []
58
0
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 SCREAMING_SNAKE_CASE_: Dict =data_utils.TransfoXLTokenizer SCREAMING_SNAKE_CASE_: Union[str, Any] =data_utils.TransfoXLCorpus SCREAMING_SNAKE_CASE_: Union[str, Any] =data_utils SCREAMING_SNAKE_CASE_: Dict =data_utils def lowerCAmelCase_ ( snake_case_ : Union[str, Any] , snake_case_ : Optional[Any] , snake_case_ : List[Any] , snake_case_ : List[Any] ) -> Any: '''simple docstring''' if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(snake_case_ , "rb" ) as fp: UpperCAmelCase_ = pickle.load(snake_case_ , encoding="latin1" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) UpperCAmelCase_ = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["pretrained_vocab_file"] print(f"""Save vocabulary to {pytorch_vocab_dump_path}""" ) UpperCAmelCase_ = corpus.vocab.__dict__ torch.save(snake_case_ , snake_case_ ) UpperCAmelCase_ = corpus.__dict__ corpus_dict_no_vocab.pop("vocab" , snake_case_ ) UpperCAmelCase_ = pytorch_dump_folder_path + "/" + CORPUS_NAME print(f"""Save dataset to {pytorch_dataset_dump_path}""" ) torch.save(snake_case_ , snake_case_ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model UpperCAmelCase_ = os.path.abspath(snake_case_ ) UpperCAmelCase_ = os.path.abspath(snake_case_ ) print(f"""Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.""" ) # Initialise PyTorch model if transfo_xl_config_file == "": UpperCAmelCase_ = TransfoXLConfig() else: UpperCAmelCase_ = TransfoXLConfig.from_json_file(snake_case_ ) print(f"""Building PyTorch model from configuration: {config}""" ) UpperCAmelCase_ = TransfoXLLMHeadModel(snake_case_ ) UpperCAmelCase_ = load_tf_weights_in_transfo_xl(snake_case_ , snake_case_ , snake_case_ ) # Save pytorch-model UpperCAmelCase_ = os.path.join(snake_case_ , snake_case_ ) UpperCAmelCase_ = os.path.join(snake_case_ , snake_case_ ) print(f"""Save PyTorch model to {os.path.abspath(snake_case_ )}""" ) torch.save(model.state_dict() , snake_case_ ) print(f"""Save configuration file to {os.path.abspath(snake_case_ )}""" ) with open(snake_case_ , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: List[Any] =argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) SCREAMING_SNAKE_CASE_: List[str] =parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
78
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = ['''image_processor''', '''tokenizer'''] _lowerCamelCase = '''BlipImageProcessor''' _lowerCamelCase = '''AutoTokenizer''' def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__(_lowercase , _lowercase ) # add QFormer tokenizer snake_case_ : List[str] = qformer_tokenizer def __call__( self , _lowercase = None , _lowercase = None , _lowercase = True , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = 0 , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = True , _lowercase = None , **_lowercase , ) -> BatchFeature: '''simple docstring''' if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) snake_case_ : Optional[Any] = BatchFeature() if text is not None: snake_case_ : List[str] = self.tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) encoding.update(_lowercase ) snake_case_ : Union[str, Any] = self.qformer_tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) snake_case_ : List[str] = qformer_text_encoding.pop("""input_ids""" ) snake_case_ : Union[str, Any] = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: snake_case_ : Tuple = self.image_processor(_lowercase , return_tensors=_lowercase ) encoding.update(_lowercase ) return encoding def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> List[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> Dict: '''simple docstring''' return self.tokenizer.decode(*_lowercase , **_lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' snake_case_ : str = self.tokenizer.model_input_names snake_case_ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def UpperCAmelCase__ ( self , _lowercase , **_lowercase ) -> Optional[int]: '''simple docstring''' if os.path.isfile(_lowercase ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(_lowercase , exist_ok=_lowercase ) snake_case_ : int = os.path.join(_lowercase , """qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(_lowercase ) return super().save_pretrained(_lowercase , **_lowercase ) @classmethod def UpperCAmelCase__ ( cls , _lowercase , **_lowercase ) -> int: '''simple docstring''' snake_case_ : List[str] = AutoTokenizer.from_pretrained(_lowercase , subfolder="""qformer_tokenizer""" ) snake_case_ : Union[str, Any] = cls._get_arguments_from_pretrained(_lowercase , **_lowercase ) args.append(_lowercase ) return cls(*_lowercase )
58
0
def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> str: '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: UpperCAmelCase__ : str = mf_knapsack(i - 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) else: UpperCAmelCase__ : str = max( mf_knapsack(i - 1 , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) , mf_knapsack(i - 1 , __lowerCamelCase , __lowerCamelCase , j - wt[i - 1] ) + val[i - 1] , ) UpperCAmelCase__ : Tuple = val return f[i][j] def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> str: '''simple docstring''' UpperCAmelCase__ : Tuple = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: UpperCAmelCase__ : Optional[Any] = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: UpperCAmelCase__ : Dict = dp[i - 1][w_] return dp[n][w_], dp def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Dict: '''simple docstring''' if not (isinstance(__lowerCamelCase , (list, tuple) ) and isinstance(__lowerCamelCase , (list, tuple) )): raise ValueError( """Both the weights and values vectors must be either lists or tuples""" ) UpperCAmelCase__ : Optional[Any] = len(__lowerCamelCase ) if num_items != len(__lowerCamelCase ): UpperCAmelCase__ : Optional[int] = ( """The number of weights must be the same as the number of values.\n""" F"But got {num_items} weights and {len(__lowerCamelCase )} values" ) raise ValueError(__lowerCamelCase ) for i in range(__lowerCamelCase ): if not isinstance(wt[i] , __lowerCamelCase ): UpperCAmelCase__ : List[str] = ( """All weights must be integers but got weight of """ F"type {type(wt[i] )} at index {i}" ) raise TypeError(__lowerCamelCase ) UpperCAmelCase__ , UpperCAmelCase__ : Any = knapsack(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) UpperCAmelCase__ : set = set() _construct_solution(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return optimal_val, example_optional_set def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> int: '''simple docstring''' # for the current item i at a maximum weight j to be part of an optimal subset, # the optimal value at (i, j) must be greater than the optimal value at (i-1, j). # where i - 1 means considering only the previous items at the given maximum weight if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(__lowerCamelCase , __lowerCamelCase , i - 1 , __lowerCamelCase , __lowerCamelCase ) else: optimal_set.add(__lowerCamelCase ) _construct_solution(__lowerCamelCase , __lowerCamelCase , i - 1 , j - wt[i - 1] , __lowerCamelCase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : List[str] = [3, 2, 4, 4] SCREAMING_SNAKE_CASE__ : Tuple = [4, 3, 2, 3] SCREAMING_SNAKE_CASE__ : Any = 4 SCREAMING_SNAKE_CASE__ : str = 6 SCREAMING_SNAKE_CASE__ : Union[str, Any] = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[int] = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : int = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("""optimal_value = """, optimal_solution) print("""An optimal subset corresponding to the optimal value""", optimal_subset)
79
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCAmelCase : List[Any] = { '''configuration_vivit''': ['''VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VivitConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = ['''VivitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = [ '''VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VivitModel''', '''VivitPreTrainedModel''', '''VivitForVideoClassification''', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
58
0
from __future__ import annotations import numpy as np def snake_case ( lowerCamelCase ): '''simple docstring''' return np.maximum(0 , lowerCamelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
80
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() __lowerCAmelCase : Dict = logging.get_logger(__name__) def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Dict , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : List[str] = WavaVecaForSequenceClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : int = downstream_dict["""projector.weight"""] snake_case_ : Optional[int] = downstream_dict["""projector.bias"""] snake_case_ : List[Any] = downstream_dict["""model.post_net.linear.weight"""] snake_case_ : Union[str, Any] = downstream_dict["""model.post_net.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Tuple , __UpperCamelCase : str ): '''simple docstring''' snake_case_ : int = WavaVecaForAudioFrameClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""model.linear.weight"""] snake_case_ : int = downstream_dict["""model.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[str] , __UpperCamelCase : List[Any] ): '''simple docstring''' snake_case_ : Optional[int] = WavaVecaForXVector.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""connector.weight"""] snake_case_ : str = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): snake_case_ : Dict = downstream_dict[ F'model.framelevel_feature_extractor.module.{i}.kernel.weight' ] snake_case_ : int = downstream_dict[F'model.framelevel_feature_extractor.module.{i}.kernel.bias'] snake_case_ : str = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] snake_case_ : Any = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] snake_case_ : List[str] = downstream_dict["""objective.W"""] return model @torch.no_grad() def __lowerCAmelCase ( __UpperCamelCase : List[Any] , __UpperCamelCase : Any , __UpperCamelCase : List[str] , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : Any = torch.load(__UpperCamelCase , map_location="""cpu""" ) snake_case_ : Any = checkpoint["""Downstream"""] snake_case_ : Optional[Any] = WavaVecaConfig.from_pretrained(__UpperCamelCase ) snake_case_ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained( __UpperCamelCase , return_attention_mask=__UpperCamelCase , do_normalize=__UpperCamelCase ) snake_case_ : Optional[Any] = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): snake_case_ : Tuple = convert_classification(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForAudioFrameClassification""" ): snake_case_ : Union[str, Any] = convert_diarization(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForXVector""" ): snake_case_ : List[str] = convert_xvector(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: raise NotImplementedError(F'S3PRL weights conversion is not supported for {arch}' ) if hf_config.use_weighted_layer_sum: snake_case_ : List[Any] = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(__UpperCamelCase ) hf_model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __lowerCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--base_model_name''', default=None, type=str, help='''Name of the huggingface pretrained base model.''' ) parser.add_argument('''--config_path''', default=None, type=str, help='''Path to the huggingface classifier config.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to the s3prl checkpoint.''') parser.add_argument('''--model_dump_path''', default=None, type=str, help='''Path to the final converted model.''') __lowerCAmelCase : Dict = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
58
0
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a : """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : Any , lowerCamelCase : Tuple=13 , lowerCamelCase : Dict=3 , lowerCamelCase : Tuple=True , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Any=0.1 , lowerCamelCase : str=0.1 , lowerCamelCase : Union[str, Any]=224 , lowerCamelCase : str=1000 , lowerCamelCase : int=[3, 3, 6, 4] , lowerCamelCase : List[str]=[48, 56, 112, 220] , ) -> List[Any]: __snake_case : Optional[int] = parent __snake_case : Tuple = batch_size __snake_case : Union[str, Any] = num_channels __snake_case : Optional[Any] = is_training __snake_case : List[str] = use_labels __snake_case : int = hidden_dropout_prob __snake_case : int = attention_probs_dropout_prob __snake_case : List[str] = num_labels __snake_case : Optional[Any] = image_size __snake_case : Dict = layer_depths __snake_case : List[Any] = embed_dims def __snake_case ( self : Optional[int] ) -> Tuple: __snake_case : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Union[str, Any] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : str = self.get_config() return config, pixel_values, labels def __snake_case ( self : List[str] ) -> Dict: return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCamelCase , layer_scale_init_value=1E-5 , ) def __snake_case ( self : int , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[int] , lowerCamelCase : List[str] ) -> List[Any]: __snake_case : Any = SwiftFormerModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Dict = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def __snake_case ( self : List[str] , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[int] ) -> List[Any]: __snake_case : Dict = self.num_labels __snake_case : Optional[Any] = SwiftFormerForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) __snake_case : List[Any] = SwiftFormerForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Any = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Optional[int] ) -> List[str]: ((__snake_case) , (__snake_case) , (__snake_case)) : List[Any] = self.prepare_config_and_inputs() __snake_case : str = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () __UpperCAmelCase : Any = ( {"feature-extraction": SwiftFormerModel, "image-classification": SwiftFormerForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase : Any = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : int = False __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : Dict = False def __snake_case ( self : Union[str, Any] ) -> Dict: __snake_case : int = SwiftFormerModelTester(self ) __snake_case : Union[str, Any] = ConfigTester( self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def __snake_case ( self : int ) -> Any: self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def __snake_case ( self : str ) -> Tuple: pass def __snake_case ( self : Optional[Any] ) -> Tuple: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : List[str] = model_class(lowerCamelCase ) __snake_case : str = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def __snake_case ( self : Dict ) -> List[str]: __snake_case , __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : int = model_class(lowerCamelCase ) __snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Optional[Any] = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> Union[str, Any]: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : int ) -> Any: __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def __snake_case ( self : Optional[int] ) -> Union[str, Any]: for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Dict = SwiftFormerModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def __snake_case ( self : Tuple ) -> Union[str, Any]: pass def __snake_case ( self : Optional[int] ) -> List[Any]: def check_hidden_states_output(lowerCamelCase : Dict , lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict ): __snake_case : Any = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : Dict = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Tuple = outputs.hidden_states __snake_case : Optional[Any] = 8 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCamelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) __snake_case , __snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Any = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : Union[str, Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Union[str, Any]: def _config_zero_init(lowerCamelCase : Optional[int] ): __snake_case : List[str] = copy.deepcopy(lowerCamelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCamelCase , lowerCamelCase , 1E-10 ) if isinstance(getattr(lowerCamelCase , lowerCamelCase , lowerCamelCase ) , lowerCamelCase ): __snake_case : Optional[Any] = _config_zero_init(getattr(lowerCamelCase , lowerCamelCase ) ) setattr(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return configs_no_init __snake_case , __snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Dict = _config_zero_init(lowerCamelCase ) for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(config=lowerCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : List[Any] ) -> Any: pass def lowerCAmelCase_ ( ): __snake_case : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : int ) -> Any: return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def __snake_case ( self : int ) -> Tuple: __snake_case : Dict = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCamelCase ) __snake_case : Dict = self.default_image_processor __snake_case : int = prepare_img() __snake_case : Tuple = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : List[str] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : Tuple = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) )
81
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __lowerCAmelCase : Tuple = logging.get_logger(__name__) __lowerCAmelCase : int = {'''vocab_file''': '''vocab.txt'''} __lowerCAmelCase : Union[str, Any] = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } __lowerCAmelCase : Optional[Any] = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } __lowerCAmelCase : Any = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = PRETRAINED_INIT_CONFIGURATION _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase = ConvBertTokenizer def __init__( self , _lowercase=None , _lowercase=None , _lowercase=True , _lowercase="[UNK]" , _lowercase="[SEP]" , _lowercase="[PAD]" , _lowercase="[CLS]" , _lowercase="[MASK]" , _lowercase=True , _lowercase=None , **_lowercase , ) -> Union[str, Any]: '''simple docstring''' super().__init__( _lowercase , tokenizer_file=_lowercase , do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , tokenize_chinese_chars=_lowercase , strip_accents=_lowercase , **_lowercase , ) snake_case_ : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowercase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowercase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowercase ) != tokenize_chinese_chars ): snake_case_ : Optional[int] = getattr(_lowercase , normalizer_state.pop("""type""" ) ) snake_case_ : Dict = do_lower_case snake_case_ : str = strip_accents snake_case_ : Optional[Any] = tokenize_chinese_chars snake_case_ : int = normalizer_class(**_lowercase ) snake_case_ : Optional[int] = do_lower_case def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> int: '''simple docstring''' snake_case_ : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> List[int]: '''simple docstring''' snake_case_ : int = [self.sep_token_id] snake_case_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[int] = self._tokenizer.model.save(_lowercase , name=_lowercase ) return tuple(_lowercase )
58
0
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np lowerCamelCase = re.compile(r"""\b(a|an|the)\b""", re.UNICODE) lowerCamelCase = None def a__ ( ): UpperCAmelCase_ = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=lowerCAmelCase__ , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=lowerCAmelCase__ , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: UpperCAmelCase_ = bool(qa["answers"]["text"] ) return qid_to_has_ans def a__ ( lowerCAmelCase__ ): def remove_articles(lowerCAmelCase__ ): return ARTICLES_REGEX.sub(" " , lowerCAmelCase__ ) def white_space_fix(lowerCAmelCase__ ): return " ".join(text.split() ) def remove_punc(lowerCAmelCase__ ): UpperCAmelCase_ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowerCAmelCase__ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowerCAmelCase__ ) ) ) ) def a__ ( lowerCAmelCase__ ): if not s: return [] return normalize_answer(lowerCAmelCase__ ).split() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): return int(normalize_answer(lowerCAmelCase__ ) == normalize_answer(lowerCAmelCase__ ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = get_tokens(lowerCAmelCase__ ) UpperCAmelCase_ = get_tokens(lowerCAmelCase__ ) UpperCAmelCase_ = collections.Counter(lowerCAmelCase__ ) & collections.Counter(lowerCAmelCase__ ) UpperCAmelCase_ = sum(common.values() ) if len(lowerCAmelCase__ ) == 0 or len(lowerCAmelCase__ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 UpperCAmelCase_ = 1.0 * num_same / len(lowerCAmelCase__ ) UpperCAmelCase_ = 1.0 * num_same / len(lowerCAmelCase__ ) UpperCAmelCase_ = (2 * precision * recall) / (precision + recall) return fa def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = {} UpperCAmelCase_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: UpperCAmelCase_ = qa["id"] UpperCAmelCase_ = [t for t in qa["answers"]["text"] if normalize_answer(lowerCAmelCase__ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string UpperCAmelCase_ = [""] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue UpperCAmelCase_ = preds[qid] # Take max over all gold answers UpperCAmelCase_ = max(compute_exact(lowerCAmelCase__ , lowerCAmelCase__ ) for a in gold_answers ) UpperCAmelCase_ = max(compute_fa(lowerCAmelCase__ , lowerCAmelCase__ ) for a in gold_answers ) return exact_scores, fa_scores def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = {} for qid, s in scores.items(): UpperCAmelCase_ = na_probs[qid] > na_prob_thresh if pred_na: UpperCAmelCase_ = float(not qid_to_has_ans[qid] ) else: UpperCAmelCase_ = s return new_scores def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): if not qid_list: UpperCAmelCase_ = len(lowerCAmelCase__ ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: UpperCAmelCase_ = len(lowerCAmelCase__ ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): for k in new_eval: UpperCAmelCase_ = new_eval[k] def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): plt.step(lowerCAmelCase__ , lowerCAmelCase__ , color="b" , alpha=0.2 , where="post" ) plt.fill_between(lowerCAmelCase__ , lowerCAmelCase__ , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(lowerCAmelCase__ ) plt.savefig(lowerCAmelCase__ ) plt.clf() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ): UpperCAmelCase_ = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : na_probs[k] ) UpperCAmelCase_ = 0.0 UpperCAmelCase_ = 1.0 UpperCAmelCase_ = 0.0 UpperCAmelCase_ = [1.0] UpperCAmelCase_ = [0.0] UpperCAmelCase_ = 0.0 for i, qid in enumerate(lowerCAmelCase__ ): if qid_to_has_ans[qid]: true_pos += scores[qid] UpperCAmelCase_ = true_pos / float(i + 1 ) UpperCAmelCase_ = true_pos / float(lowerCAmelCase__ ) if i == len(lowerCAmelCase__ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(lowerCAmelCase__ ) recalls.append(lowerCAmelCase__ ) if out_image: plot_pr_curve(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return {"ap": 100.0 * avg_prec} def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): if out_image_dir and not os.path.exists(lowerCAmelCase__ ): os.makedirs(lowerCAmelCase__ ) UpperCAmelCase_ = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return UpperCAmelCase_ = make_precision_recall_eval( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , out_image=os.path.join(lowerCAmelCase__ , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) UpperCAmelCase_ = make_precision_recall_eval( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , out_image=os.path.join(lowerCAmelCase__ , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) UpperCAmelCase_ = {k: float(lowerCAmelCase__ ) for k, v in qid_to_has_ans.items()} UpperCAmelCase_ = make_precision_recall_eval( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , out_image=os.path.join(lowerCAmelCase__ , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "pr_exact" ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "pr_f1" ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "pr_oracle" ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): if not qid_list: return UpperCAmelCase_ = [na_probs[k] for k in qid_list] UpperCAmelCase_ = np.ones_like(lowerCAmelCase__ ) / float(len(lowerCAmelCase__ ) ) plt.hist(lowerCAmelCase__ , weights=lowerCAmelCase__ , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(lowerCAmelCase__ , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) UpperCAmelCase_ = num_no_ans UpperCAmelCase_ = cur_score UpperCAmelCase_ = 0.0 UpperCAmelCase_ = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : na_probs[k] ) for i, qid in enumerate(lowerCAmelCase__ ): if qid not in scores: continue if qid_to_has_ans[qid]: UpperCAmelCase_ = scores[qid] else: if preds[qid]: UpperCAmelCase_ = -1 else: UpperCAmelCase_ = 0 cur_score += diff if cur_score > best_score: UpperCAmelCase_ = cur_score UpperCAmelCase_ = na_probs[qid] return 100.0 * best_score / len(lowerCAmelCase__ ), best_thresh def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ , UpperCAmelCase_ = find_best_thresh(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ , UpperCAmelCase_ = find_best_thresh(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = best_exact UpperCAmelCase_ = exact_thresh UpperCAmelCase_ = best_fa UpperCAmelCase_ = fa_thresh def a__ ( ): with open(OPTS.data_file ) as f: UpperCAmelCase_ = json.load(lowerCAmelCase__ ) UpperCAmelCase_ = dataset_json["data"] with open(OPTS.pred_file ) as f: UpperCAmelCase_ = json.load(lowerCAmelCase__ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: UpperCAmelCase_ = json.load(lowerCAmelCase__ ) else: UpperCAmelCase_ = {k: 0.0 for k in preds} UpperCAmelCase_ = make_qid_to_has_ans(lowerCAmelCase__ ) # maps qid to True/False UpperCAmelCase_ = [k for k, v in qid_to_has_ans.items() if v] UpperCAmelCase_ = [k for k, v in qid_to_has_ans.items() if not v] UpperCAmelCase_ , UpperCAmelCase_ = get_raw_scores(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = apply_no_ans_threshold(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , OPTS.na_prob_thresh ) UpperCAmelCase_ = apply_no_ans_threshold(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , OPTS.na_prob_thresh ) UpperCAmelCase_ = make_eval_dict(lowerCAmelCase__ , lowerCAmelCase__ ) if has_ans_qids: UpperCAmelCase_ = make_eval_dict(lowerCAmelCase__ , lowerCAmelCase__ , qid_list=lowerCAmelCase__ ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "HasAns" ) if no_ans_qids: UpperCAmelCase_ = make_eval_dict(lowerCAmelCase__ , lowerCAmelCase__ , qid_list=lowerCAmelCase__ ) merge_eval(lowerCAmelCase__ , lowerCAmelCase__ , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , OPTS.out_image_dir ) histogram_na_prob(lowerCAmelCase__ , lowerCAmelCase__ , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(lowerCAmelCase__ , lowerCAmelCase__ , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) else: print(json.dumps(lowerCAmelCase__ , indent=2 ) ) if __name__ == "__main__": lowerCamelCase = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("""Agg""") import matplotlib.pyplot as plt main()
82
"""simple docstring""" import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" @register_to_config def __init__( self , _lowercase = 1_2_8 , _lowercase = 2_5_6 , _lowercase = 2000.0 , _lowercase = 7_6_8 , _lowercase = 1_2 , _lowercase = 1_2 , _lowercase = 6_4 , _lowercase = 2_0_4_8 , _lowercase = 0.1 , ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Optional[Any] = nn.Sequential( nn.Linear(_lowercase , d_model * 4 , bias=_lowercase ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_lowercase ) , nn.SiLU() , ) snake_case_ : Any = nn.Embedding(_lowercase , _lowercase ) snake_case_ : Union[str, Any] = False snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Union[str, Any] = nn.Dropout(p=_lowercase ) snake_case_ : Tuple = nn.ModuleList() for lyr_num in range(_lowercase ): # FiLM conditional T5 decoder snake_case_ : Union[str, Any] = DecoderLayer(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) self.decoders.append(_lowercase ) snake_case_ : List[Any] = TaLayerNorm(_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(p=_lowercase ) snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def UpperCAmelCase__ ( self , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ , snake_case_ , snake_case_ : str = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. snake_case_ : Optional[int] = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) snake_case_ : int = self.conditioning_emb(_lowercase ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) snake_case_ : Tuple = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. snake_case_ : Dict = torch.broadcast_to( torch.arange(_lowercase , device=decoder_input_tokens.device ) , (batch, seq_length) , ) snake_case_ : Tuple = self.position_encoding(_lowercase ) snake_case_ : Optional[Any] = self.continuous_inputs_projection(_lowercase ) inputs += position_encodings snake_case_ : List[Any] = self.dropout(_lowercase ) # decoder: No padding present. snake_case_ : Tuple = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. snake_case_ : int = [(x, self.encoder_decoder_mask(_lowercase , _lowercase )) for x, y in encodings_and_masks] # cross attend style: concat encodings snake_case_ : Optional[Any] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) snake_case_ : str = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: snake_case_ : int = lyr( _lowercase , conditioning_emb=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , )[0] snake_case_ : int = self.decoder_norm(_lowercase ) snake_case_ : Union[str, Any] = self.post_dropout(_lowercase ) snake_case_ : int = self.spec_out(_lowercase ) return spec_out class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase=1E-6 ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Any = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase ) ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = self.layer[0]( _lowercase , conditioning_emb=_lowercase , attention_mask=_lowercase , ) if encoder_hidden_states is not None: snake_case_ : Tuple = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) snake_case_ : str = self.layer[1]( _lowercase , key_value_states=_lowercase , attention_mask=_lowercase , ) # Apply Film Conditional Feed Forward layer snake_case_ : Any = self.layer[-1](_lowercase , _lowercase ) return (hidden_states,) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> str: '''simple docstring''' super().__init__() snake_case_ : Any = TaLayerNorm(_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Union[str, Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : List[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : str = self.FiLMLayer(_lowercase , _lowercase ) # Self-attention block snake_case_ : List[Any] = self.attention(_lowercase ) snake_case_ : List[str] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__() snake_case_ : List[Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : Union[str, Any] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) snake_case_ : Optional[Any] = self.attention( _lowercase , encoder_hidden_states=_lowercase , attention_mask=attention_mask.squeeze(1 ) , ) snake_case_ : Any = hidden_states + self.dropout(_lowercase ) return layer_output class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Tuple = TaDenseGatedActDense(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Optional[int] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Tuple = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> str: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : Optional[int] = self.film(_lowercase , _lowercase ) snake_case_ : int = self.DenseReluDense(_lowercase ) snake_case_ : Optional[Any] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__() snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Any = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : int = nn.Dropout(_lowercase ) snake_case_ : Optional[int] = NewGELUActivation() def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' snake_case_ : str = self.act(self.wi_a(_lowercase ) ) snake_case_ : Dict = self.wi_a(_lowercase ) snake_case_ : Any = hidden_gelu * hidden_linear snake_case_ : List[Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.wo(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1E-6 ) -> str: '''simple docstring''' super().__init__() snake_case_ : Union[str, Any] = nn.Parameter(torch.ones(_lowercase ) ) snake_case_ : int = eps def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=_lowercase ) snake_case_ : Any = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: snake_case_ : str = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def UpperCAmelCase__ ( self , _lowercase ) -> torch.Tensor: '''simple docstring''' return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.04_4715 * torch.pow(_lowercase , 3.0 )) )) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Any: '''simple docstring''' super().__init__() snake_case_ : List[Any] = nn.Linear(_lowercase , out_features * 2 , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = self.scale_bias(_lowercase ) snake_case_ , snake_case_ : Any = torch.chunk(_lowercase , 2 , -1 ) snake_case_ : Optional[Any] = x * (1 + scale) + shift return x
58
0
"""simple docstring""" import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def snake_case_ ( A_ : Optional[int] ): '''simple docstring''' _lowerCamelCase : int = SwinConfig() _lowerCamelCase : Dict = swin_name.split('''_''' ) _lowerCamelCase : List[Any] = name_split[1] _lowerCamelCase : List[Any] = int(name_split[4] ) _lowerCamelCase : str = int(name_split[3][-1] ) if model_size == "tiny": _lowerCamelCase : Dict = 96 _lowerCamelCase : Tuple = (2, 2, 6, 2) _lowerCamelCase : List[Any] = (3, 6, 12, 24) elif model_size == "small": _lowerCamelCase : Union[str, Any] = 96 _lowerCamelCase : List[Any] = (2, 2, 18, 2) _lowerCamelCase : str = (3, 6, 12, 24) elif model_size == "base": _lowerCamelCase : int = 1_28 _lowerCamelCase : List[str] = (2, 2, 18, 2) _lowerCamelCase : Tuple = (4, 8, 16, 32) else: _lowerCamelCase : Optional[int] = 1_92 _lowerCamelCase : Any = (2, 2, 18, 2) _lowerCamelCase : Tuple = (6, 12, 24, 48) if "in22k" in swin_name: _lowerCamelCase : Optional[int] = 2_18_41 else: _lowerCamelCase : Optional[int] = 10_00 _lowerCamelCase : str = '''huggingface/label-files''' _lowerCamelCase : str = '''imagenet-1k-id2label.json''' _lowerCamelCase : Any = json.load(open(hf_hub_download(A_, A_, repo_type='''dataset''' ), '''r''' ) ) _lowerCamelCase : Optional[int] = {int(A_ ): v for k, v in idalabel.items()} _lowerCamelCase : List[str] = idalabel _lowerCamelCase : int = {v: k for k, v in idalabel.items()} _lowerCamelCase : List[Any] = img_size _lowerCamelCase : int = num_classes _lowerCamelCase : Optional[int] = embed_dim _lowerCamelCase : Optional[int] = depths _lowerCamelCase : Optional[Any] = num_heads _lowerCamelCase : Union[str, Any] = window_size return config def snake_case_ ( A_ : int ): '''simple docstring''' if "patch_embed.proj" in name: _lowerCamelCase : Optional[int] = name.replace('''patch_embed.proj''', '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: _lowerCamelCase : Optional[int] = name.replace('''patch_embed.norm''', '''embeddings.norm''' ) if "layers" in name: _lowerCamelCase : str = '''encoder.''' + name if "attn.proj" in name: _lowerCamelCase : List[Any] = name.replace('''attn.proj''', '''attention.output.dense''' ) if "attn" in name: _lowerCamelCase : Dict = name.replace('''attn''', '''attention.self''' ) if "norm1" in name: _lowerCamelCase : List[Any] = name.replace('''norm1''', '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase : List[Any] = name.replace('''norm2''', '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase : List[str] = name.replace('''mlp.fc1''', '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase : Optional[int] = name.replace('''mlp.fc2''', '''output.dense''' ) if name == "norm.weight": _lowerCamelCase : Any = '''layernorm.weight''' if name == "norm.bias": _lowerCamelCase : Union[str, Any] = '''layernorm.bias''' if "head" in name: _lowerCamelCase : Dict = name.replace('''head''', '''classifier''' ) else: _lowerCamelCase : List[Any] = '''swin.''' + name return name def snake_case_ ( A_ : List[str], A_ : int ): '''simple docstring''' for key in orig_state_dict.copy().keys(): _lowerCamelCase : int = orig_state_dict.pop(A_ ) if "mask" in key: continue elif "qkv" in key: _lowerCamelCase : Optional[int] = key.split('''.''' ) _lowerCamelCase : Dict = int(key_split[1] ) _lowerCamelCase : Dict = int(key_split[3] ) _lowerCamelCase : List[Any] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase : Optional[Any] = val[:dim, :] _lowerCamelCase : int = val[ dim : dim * 2, : ] _lowerCamelCase : Optional[Any] = val[-dim:, :] else: _lowerCamelCase : Optional[int] = val[ :dim ] _lowerCamelCase : int = val[ dim : dim * 2 ] _lowerCamelCase : Union[str, Any] = val[ -dim: ] else: _lowerCamelCase : List[Any] = val return orig_state_dict def snake_case_ ( A_ : Any, A_ : Tuple ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = timm.create_model(A_, pretrained=A_ ) timm_model.eval() _lowerCamelCase : Union[str, Any] = get_swin_config(A_ ) _lowerCamelCase : Union[str, Any] = SwinForImageClassification(A_ ) model.eval() _lowerCamelCase : Tuple = convert_state_dict(timm_model.state_dict(), A_ ) model.load_state_dict(A_ ) _lowerCamelCase : Dict = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase : Optional[int] = AutoImageProcessor.from_pretrained('''microsoft/{}'''.format(swin_name.replace('''_''', '''-''' ) ) ) _lowerCamelCase : str = Image.open(requests.get(A_, stream=A_ ).raw ) _lowerCamelCase : Tuple = image_processor(images=A_, return_tensors='''pt''' ) _lowerCamelCase : Dict = timm_model(inputs['''pixel_values'''] ) _lowerCamelCase : Union[str, Any] = model(**A_ ).logits assert torch.allclose(A_, A_, atol=1E-3 ) print(F'''Saving model {swin_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(A_ ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(A_ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swin_name''', default='''swin_tiny_patch4_window7_224''', type=str, help='''Name of the Swin timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase__ = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
83
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''roformer''' def __init__( self , _lowercase=5_0_0_0_0 , _lowercase=None , _lowercase=7_6_8 , _lowercase=1_2 , _lowercase=1_2 , _lowercase=3_0_7_2 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=1_5_3_6 , _lowercase=2 , _lowercase=0.02 , _lowercase=1E-12 , _lowercase=0 , _lowercase=False , _lowercase=True , **_lowercase , ) -> int: '''simple docstring''' super().__init__(pad_token_id=_lowercase , **_lowercase ) snake_case_ : str = vocab_size snake_case_ : Any = hidden_size if embedding_size is None else embedding_size snake_case_ : List[str] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : Tuple = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : List[str] = rotary_value snake_case_ : str = use_cache class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ : Any = {0: """batch""", 1: """sequence"""} snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
58
0
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract UpperCAmelCase = logging.get_logger(__name__) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return [ int(1000 * (box[0] / width) ), int(1000 * (box[1] / height) ), int(1000 * (box[2] / width) ), int(1000 * (box[3] / height) ), ] def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None ): lowercase = tesseract_config if tesseract_config is not None else '' # apply OCR lowercase = to_pil_image(__SCREAMING_SNAKE_CASE ) lowercase , lowercase = pil_image.size lowercase = pytesseract.image_to_data(__SCREAMING_SNAKE_CASE , lang=__SCREAMING_SNAKE_CASE , output_type='dict' , config=__SCREAMING_SNAKE_CASE ) lowercase , lowercase , lowercase , lowercase , lowercase = data['text'], data['left'], data['top'], data['width'], data['height'] # filter empty words and corresponding coordinates lowercase = [idx for idx, word in enumerate(__SCREAMING_SNAKE_CASE ) if not word.strip()] lowercase = [word for idx, word in enumerate(__SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] lowercase = [coord for idx, coord in enumerate(__SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] lowercase = [coord for idx, coord in enumerate(__SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] lowercase = [coord for idx, coord in enumerate(__SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] lowercase = [coord for idx, coord in enumerate(__SCREAMING_SNAKE_CASE ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format lowercase = [] for x, y, w, h in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = [x, y, x + w, y + h] actual_boxes.append(__SCREAMING_SNAKE_CASE ) # finally, normalize the bounding boxes lowercase = [] for box in actual_boxes: normalized_boxes.append(normalize_box(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) assert len(__SCREAMING_SNAKE_CASE ) == len(__SCREAMING_SNAKE_CASE ), "Not as many words as there are bounding boxes" return words, normalized_boxes class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[str] = ["""pixel_values"""] def __init__( self , snake_case = True , snake_case = None , snake_case = PILImageResampling.BILINEAR , snake_case = True , snake_case = None , snake_case = "" , **snake_case , ): super().__init__(**snake_case ) lowercase = size if size is not None else {'height': 224, 'width': 224} lowercase = get_size_dict(snake_case ) lowercase = do_resize lowercase = size lowercase = resample lowercase = apply_ocr lowercase = ocr_lang lowercase = tesseract_config def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case = PILImageResampling.BILINEAR , snake_case = None , **snake_case , ): lowercase = get_size_dict(snake_case ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) lowercase = (size['height'], size['width']) return resize(snake_case , size=snake_case , resample=snake_case , data_format=snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = ChannelDimension.FIRST , **snake_case , ): lowercase = do_resize if do_resize is not None else self.do_resize lowercase = size if size is not None else self.size lowercase = get_size_dict(snake_case ) lowercase = resample if resample is not None else self.resample lowercase = apply_ocr if apply_ocr is not None else self.apply_ocr lowercase = ocr_lang if ocr_lang is not None else self.ocr_lang lowercase = tesseract_config if tesseract_config is not None else self.tesseract_config lowercase = make_list_of_images(snake_case ) if not valid_images(snake_case ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) # All transformations expect numpy arrays. lowercase = [to_numpy_array(snake_case ) for image in images] if apply_ocr: requires_backends(self , 'pytesseract' ) lowercase = [] lowercase = [] for image in images: lowercase , lowercase = apply_tesseract(snake_case , snake_case , snake_case ) words_batch.append(snake_case ) boxes_batch.append(snake_case ) if do_resize: lowercase = [self.resize(image=snake_case , size=snake_case , resample=snake_case ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) lowercase = [flip_channel_order(snake_case ) for image in images] lowercase = [to_channel_dimension_format(snake_case , snake_case ) for image in images] lowercase = BatchFeature(data={'pixel_values': images} , tensor_type=snake_case ) if apply_ocr: lowercase = words_batch lowercase = boxes_batch return data
84
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __lowerCAmelCase ( __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Dict = checkpoints.load_tax_checkpoint(__UpperCamelCase ) snake_case_ : Tuple = flatten_dict(__UpperCamelCase ) return flax_params def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] ): '''simple docstring''' snake_case_ : Optional[Any] = {} snake_case_ : List[Any] = { """token_embedder""": """embeddings""", """encoder_norm""": """layernorm""", """kernel""": """weight""", """.out""": """.output""", """scale""": """weight""", """embedders_0.pos_embedding""": """row_embedder.weight""", """embedders_1.pos_embedding""": """column_embedder.weight""", } snake_case_ : Optional[Any] = { """query""": """attention.query""", """key""": """attention.key""", """value""": """attention.value""", """output.dense""": """output""", """encoder_decoder_attention.o""": """encoder_decoder_attention.attention.o""", """pre_self_attention_layer_norm""": """self_attention.layer_norm""", """pre_cross_attention_layer_norm""": """encoder_decoder_attention.layer_norm""", """mlp.""": """mlp.DenseReluDense.""", """pre_mlp_layer_norm""": """mlp.layer_norm""", """self_attention.o""": """self_attention.attention.o""", """decoder.embeddings.embedding""": """decoder.embed_tokens.weight""", """decoder.relpos_bias.rel_embedding""": """decoder.layer.0.self_attention.attention.relative_attention_bias.weight""", """decoder.decoder_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.logits_dense.weight""": """decoder.lm_head.weight""", } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key snake_case_ : List[Any] = """.""".join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): snake_case_ : List[str] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): snake_case_ : Optional[int] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number snake_case_ : Optional[Any] = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Union[str, Any] = new_key.replace("""encoder""" , """encoder.encoder""" ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number snake_case_ : int = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Dict = flax_dict[key] snake_case_ : Tuple = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): snake_case_ : Optional[int] = torch.from_numpy(converted_dict[key].T ) else: snake_case_ : List[Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : List[str]=False ): '''simple docstring''' snake_case_ : Optional[int] = get_flax_param(__UpperCamelCase ) if not use_large: snake_case_ : Optional[int] = PixaStructVisionConfig() snake_case_ : Optional[Any] = PixaStructTextConfig() else: snake_case_ : Tuple = PixaStructVisionConfig( hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_attention_heads=2_4 , num_hidden_layers=1_8 ) snake_case_ : List[str] = PixaStructTextConfig(hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_heads=2_4 , num_layers=1_8 ) snake_case_ : str = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=__UpperCamelCase ) snake_case_ : Optional[int] = PixaStructForConditionalGeneration(__UpperCamelCase ) snake_case_ : str = rename_and_convert_flax_params(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) snake_case_ : Optional[int] = AutoTokenizer.from_pretrained("""ybelkada/test-pix2struct-tokenizer""" ) snake_case_ : int = PixaStructImageProcessor() snake_case_ : str = PixaStructProcessor(image_processor=__UpperCamelCase , tokenizer=__UpperCamelCase ) if use_large: snake_case_ : Optional[Any] = 4_0_9_6 snake_case_ : int = True # mkdir if needed os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) processor.save_pretrained(__UpperCamelCase ) print("""Model saved in {}""".format(__UpperCamelCase ) ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') __lowerCAmelCase : List[Any] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
58
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE__ : List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Optional[Any] = { "ut/deta": "https://huggingface.co/ut/deta/resolve/main/config.json", } class snake_case ( UpperCamelCase_ ): lowercase_ = 'deta' lowercase_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Union[str, Any] , a_ : Dict=None , a_ : Tuple=900 , a_ : Any=2048 , a_ : List[str]=6 , a_ : int=2048 , a_ : Union[str, Any]=8 , a_ : List[Any]=6 , a_ : List[Any]=1024 , a_ : Union[str, Any]=8 , a_ : List[Any]=0.0 , a_ : List[Any]=True , a_ : str="relu" , a_ : Any=256 , a_ : Optional[Any]=0.1 , a_ : Dict=0.0 , a_ : Union[str, Any]=0.0 , a_ : Optional[int]=0.02 , a_ : Optional[Any]=1.0 , a_ : Dict=True , a_ : int=False , a_ : List[str]="sine" , a_ : Dict=5 , a_ : Tuple=4 , a_ : Union[str, Any]=4 , a_ : Dict=True , a_ : str=300 , a_ : Union[str, Any]=True , a_ : List[Any]=True , a_ : List[Any]=1 , a_ : List[str]=5 , a_ : Optional[int]=2 , a_ : List[str]=1 , a_ : Dict=1 , a_ : List[str]=5 , a_ : List[Any]=2 , a_ : Union[str, Any]=0.1 , a_ : int=0.25 , **a_ : List[str] , )-> int: """simple docstring""" if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) SCREAMING_SNAKE_CASE__ : Optional[int] = CONFIG_MAPPING['resnet'](out_features=['stage2', 'stage3', 'stage4'] ) else: if isinstance(a_ , a_ ): SCREAMING_SNAKE_CASE__ : Optional[int] = backbone_config.pop('model_type' ) SCREAMING_SNAKE_CASE__ : Dict = CONFIG_MAPPING[backbone_model_type] SCREAMING_SNAKE_CASE__ : List[Any] = config_class.from_dict(a_ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = backbone_config SCREAMING_SNAKE_CASE__ : Any = num_queries SCREAMING_SNAKE_CASE__ : Optional[Any] = max_position_embeddings SCREAMING_SNAKE_CASE__ : Tuple = d_model SCREAMING_SNAKE_CASE__ : Optional[Any] = encoder_ffn_dim SCREAMING_SNAKE_CASE__ : Union[str, Any] = encoder_layers SCREAMING_SNAKE_CASE__ : str = encoder_attention_heads SCREAMING_SNAKE_CASE__ : str = decoder_ffn_dim SCREAMING_SNAKE_CASE__ : List[Any] = decoder_layers SCREAMING_SNAKE_CASE__ : List[Any] = decoder_attention_heads SCREAMING_SNAKE_CASE__ : str = dropout SCREAMING_SNAKE_CASE__ : Dict = attention_dropout SCREAMING_SNAKE_CASE__ : Optional[int] = activation_dropout SCREAMING_SNAKE_CASE__ : int = activation_function SCREAMING_SNAKE_CASE__ : List[Any] = init_std SCREAMING_SNAKE_CASE__ : List[Any] = init_xavier_std SCREAMING_SNAKE_CASE__ : str = encoder_layerdrop SCREAMING_SNAKE_CASE__ : List[str] = auxiliary_loss SCREAMING_SNAKE_CASE__ : Tuple = position_embedding_type # deformable attributes SCREAMING_SNAKE_CASE__ : List[Any] = num_feature_levels SCREAMING_SNAKE_CASE__ : Optional[int] = encoder_n_points SCREAMING_SNAKE_CASE__ : Optional[Any] = decoder_n_points SCREAMING_SNAKE_CASE__ : Any = two_stage SCREAMING_SNAKE_CASE__ : Union[str, Any] = two_stage_num_proposals SCREAMING_SNAKE_CASE__ : Any = with_box_refine SCREAMING_SNAKE_CASE__ : Union[str, Any] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.' ) # Hungarian matcher SCREAMING_SNAKE_CASE__ : Dict = class_cost SCREAMING_SNAKE_CASE__ : Optional[int] = bbox_cost SCREAMING_SNAKE_CASE__ : int = giou_cost # Loss coefficients SCREAMING_SNAKE_CASE__ : Any = mask_loss_coefficient SCREAMING_SNAKE_CASE__ : Optional[int] = dice_loss_coefficient SCREAMING_SNAKE_CASE__ : int = bbox_loss_coefficient SCREAMING_SNAKE_CASE__ : Optional[int] = giou_loss_coefficient SCREAMING_SNAKE_CASE__ : str = eos_coefficient SCREAMING_SNAKE_CASE__ : List[str] = focal_alpha super().__init__(is_encoder_decoder=a_ , **a_ ) @property def __lowercase( self : Optional[int] )-> int: """simple docstring""" return self.encoder_attention_heads @property def __lowercase( self : Optional[Any] )-> int: """simple docstring""" return self.d_model def __lowercase( self : Optional[int] )-> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ : Tuple = self.backbone_config.to_dict() SCREAMING_SNAKE_CASE__ : Dict = self.__class__.model_type return output
85
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if mass < 0: raise ValueError("""The mass of a body cannot be negative""" ) return 0.5 * mass * abs(__UpperCamelCase ) * abs(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
58
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() __a :Dict = logging.get_logger(__name__) def __snake_case ( __UpperCamelCase : Dict ,__UpperCamelCase : Tuple=False ): """simple docstring""" A_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" A_ = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def __snake_case ( __UpperCamelCase : Optional[int] ,__UpperCamelCase : Tuple ,__UpperCamelCase : Any=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: A_ = "" else: A_ = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A_ = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) A_ = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict A_ = in_proj_weight[ : config.hidden_size, : ] A_ = in_proj_bias[: config.hidden_size] A_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A_ = in_proj_weight[ -config.hidden_size :, : ] A_ = in_proj_bias[-config.hidden_size :] def __snake_case ( __UpperCamelCase : List[Any] ): """simple docstring""" A_ = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(__UpperCamelCase ,__UpperCamelCase ) def __snake_case ( __UpperCamelCase : Any ,__UpperCamelCase : Optional[int] ,__UpperCamelCase : List[str] ): """simple docstring""" A_ = dct.pop(__UpperCamelCase ) A_ = val def __snake_case ( ): """simple docstring""" A_ = "http://images.cocodataset.org/val2017/000000039769.jpg" A_ = Image.open(requests.get(__UpperCamelCase ,stream=__UpperCamelCase ).raw ) return im @torch.no_grad() def __snake_case ( __UpperCamelCase : Union[str, Any] ,__UpperCamelCase : Optional[int] ): """simple docstring""" A_ = ViTConfig() A_ = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": A_ = True A_ = int(vit_name[-12:-10] ) A_ = int(vit_name[-9:-6] ) else: A_ = 1000 A_ = "huggingface/label-files" A_ = "imagenet-1k-id2label.json" A_ = json.load(open(hf_hub_download(__UpperCamelCase ,__UpperCamelCase ,repo_type="dataset" ) ,"r" ) ) A_ = {int(__UpperCamelCase ): v for k, v in idalabel.items()} A_ = idalabel A_ = {v: k for k, v in idalabel.items()} A_ = int(vit_name[-6:-4] ) A_ = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith("tiny" ): A_ = 192 A_ = 768 A_ = 12 A_ = 3 elif vit_name[9:].startswith("small" ): A_ = 384 A_ = 1536 A_ = 12 A_ = 6 else: pass else: if vit_name[4:].startswith("small" ): A_ = 768 A_ = 2304 A_ = 8 A_ = 8 elif vit_name[4:].startswith("base" ): pass elif vit_name[4:].startswith("large" ): A_ = 1024 A_ = 4096 A_ = 24 A_ = 16 elif vit_name[4:].startswith("huge" ): A_ = 1280 A_ = 5120 A_ = 32 A_ = 16 # load original model from timm A_ = timm.create_model(__UpperCamelCase ,pretrained=__UpperCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys A_ = timm_model.state_dict() if base_model: remove_classification_head_(__UpperCamelCase ) A_ = create_rename_keys(__UpperCamelCase ,__UpperCamelCase ) for src, dest in rename_keys: rename_key(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) read_in_q_k_v(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # load HuggingFace model if vit_name[-5:] == "in21k": A_ = ViTModel(__UpperCamelCase ).eval() else: A_ = ViTForImageClassification(__UpperCamelCase ).eval() model.load_state_dict(__UpperCamelCase ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: A_ = DeiTImageProcessor(size=config.image_size ) else: A_ = ViTImageProcessor(size=config.image_size ) A_ = image_processor(images=prepare_img() ,return_tensors="pt" ) A_ = encoding["pixel_values"] A_ = model(__UpperCamelCase ) if base_model: A_ = timm_model.forward_features(__UpperCamelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__UpperCamelCase ,outputs.pooler_output ,atol=1E-3 ) else: A_ = timm_model(__UpperCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__UpperCamelCase ,outputs.logits ,atol=1E-3 ) Path(__UpperCamelCase ).mkdir(exist_ok=__UpperCamelCase ) print(f'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__UpperCamelCase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __a :str = argparse.ArgumentParser() # Required parameters parser.add_argument( '--vit_name', default='vit_base_patch16_224', type=str, help='Name of the ViT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) __a :Optional[int] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
86
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = StableDiffusionInpaintPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase = frozenset([] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=_lowercase , ) snake_case_ : Dict = PNDMScheduler(skip_prk_steps=_lowercase ) torch.manual_seed(0 ) snake_case_ : str = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) snake_case_ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) snake_case_ : Dict = CLIPTextModel(_lowercase ) snake_case_ : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case_ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self , _lowercase , _lowercase=0 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowercase ) ).to(_lowercase ) snake_case_ : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case_ : Tuple = Image.fromarray(np.uinta(_lowercase ) ).convert("""RGB""" ).resize((6_4, 6_4) ) snake_case_ : Any = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((6_4, 6_4) ) if str(_lowercase ).startswith("""mps""" ): snake_case_ : str = torch.manual_seed(_lowercase ) else: snake_case_ : List[str] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) snake_case_ : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case_ : List[str] = self.get_dummy_components() snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline(**_lowercase ) snake_case_ : Dict = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) snake_case_ : Optional[int] = self.get_dummy_inputs(_lowercase ) snake_case_ : List[str] = sd_pipe(**_lowercase ).images snake_case_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) snake_case_ : Optional[int] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained(_lowercase , safety_checker=_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : List[Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[str] = torch.manual_seed(0 ) snake_case_ : Dict = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ : str = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , torch_dtype=torch.floataa , safety_checker=_lowercase , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : Union[str, Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : Optional[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : str = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case_ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : int = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : List[str] = PNDMScheduler.from_pretrained(_lowercase , subfolder="""scheduler""" ) snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , safety_checker=_lowercase , scheduler=_lowercase , torch_dtype=torch.floataa , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case_ : Any = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , num_inference_steps=2 , output_type="""np""" , ) snake_case_ : Dict = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
58
0
def SCREAMING_SNAKE_CASE ( lowercase_ ) -> list: """simple docstring""" if len(lowercase_ ) <= 1: return [tuple(lowercase_ )] A__ = [] def generate(lowercase_ , lowercase_ ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , lowercase_ ) for i in range(k - 1 ): if k % 2 == 0: # k is even A__ , A__ = arr[k - 1], arr[i] else: # k is odd A__ , A__ = arr[k - 1], arr[0] generate(k - 1 , lowercase_ ) generate(len(lowercase_ ) , lowercase_ ) return res if __name__ == "__main__": _lowerCamelCase : int = input("""Enter numbers separated by a comma:\n""").strip() _lowerCamelCase : str = [int(item) for item in user_input.split(""",""")] print(heaps(arr))
87
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : str , __UpperCamelCase : Any ): '''simple docstring''' snake_case_ : Optional[Any] = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] snake_case_ : Optional[int] = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } snake_case_ : Optional[Any] = F'{src_lang}-{tgt_lang}' snake_case_ : Dict = F'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) snake_case_ : List[str] = os.path.join(__UpperCamelCase , """README.md""" ) print(F'Generating {path}' ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(__UpperCamelCase ) # make sure we are under the root of the project __lowerCAmelCase : str = Path(__file__).resolve().parent.parent.parent __lowerCAmelCase : Optional[int] = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[int] = model_name.split('''-''') __lowerCAmelCase : Optional[int] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
58
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCAmelCase = { """configuration_vision_encoder_decoder""": ["""VisionEncoderDecoderConfig""", """VisionEncoderDecoderOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["""VisionEncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["""TFVisionEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ["""FlaxVisionEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
88
"""simple docstring""" __lowerCAmelCase : Tuple = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowerCAmelCase : Tuple = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __lowerCAmelCase : Any = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
58
0
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class _lowerCamelCase( _a ): def __init__( self, *lowerCamelCase, lowerCamelCase=None, lowerCamelCase=None, **lowerCamelCase) -> List[Any]: """simple docstring""" super().__init__(*lowerCamelCase, **lowerCamelCase) _lowercase : Any = eval_examples _lowercase : List[Any] = post_process_function def UpperCamelCase ( self, lowerCamelCase = None, lowerCamelCase=None, lowerCamelCase = None, lowerCamelCase = "eval", **lowerCamelCase, ) -> Dict[str, float]: """simple docstring""" _lowercase : Optional[Any] = gen_kwargs.copy() _lowercase : List[Any] = ( gen_kwargs['max_length'] if gen_kwargs.get('max_length') is not None else self.args.generation_max_length ) _lowercase : Any = ( gen_kwargs['num_beams'] if gen_kwargs.get('num_beams') is not None else self.args.generation_num_beams ) _lowercase : Optional[Any] = gen_kwargs _lowercase : Optional[int] = self.eval_dataset if eval_dataset is None else eval_dataset _lowercase : Optional[int] = self.get_eval_dataloader(lowerCamelCase) _lowercase : Optional[Any] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. _lowercase : List[Any] = self.compute_metrics _lowercase : int = None _lowercase : Optional[Any] = time.time() _lowercase : Optional[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _lowercase : int = eval_loop( lowerCamelCase, description='Evaluation', prediction_loss_only=True if compute_metrics is None else None, ignore_keys=lowerCamelCase, metric_key_prefix=lowerCamelCase, ) finally: _lowercase : Union[str, Any] = compute_metrics _lowercase : Optional[int] = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( lowerCamelCase, lowerCamelCase, num_samples=output.num_samples, num_steps=math.ceil(output.num_samples / total_batch_size), )) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default _lowercase : Dict = self.post_process_function(lowerCamelCase, lowerCamelCase, lowerCamelCase) _lowercase : List[Any] = self.compute_metrics(lowerCamelCase) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(F'''{metric_key_prefix}_'''): _lowercase : Optional[int] = metrics.pop(lowerCamelCase) metrics.update(output.metrics) else: _lowercase : Dict = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(lowerCamelCase) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) _lowercase : List[str] = self.callback_handler.on_evaluate(self.args, self.state, self.control, lowerCamelCase) return metrics def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase=None, lowerCamelCase = "test", **lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : str = gen_kwargs.copy() _lowercase : str = self.get_test_dataloader(lowerCamelCase) # Temporarily disable metric computation, we will do it in the loop here. _lowercase : List[str] = self.compute_metrics _lowercase : Any = None _lowercase : str = time.time() _lowercase : Optional[int] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: _lowercase : str = eval_loop( lowerCamelCase, description='Prediction', prediction_loss_only=True if compute_metrics is None else None, ignore_keys=lowerCamelCase, metric_key_prefix=lowerCamelCase, ) finally: _lowercase : int = compute_metrics _lowercase : int = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( lowerCamelCase, lowerCamelCase, num_samples=output.num_samples, num_steps=math.ceil(output.num_samples / total_batch_size), )) if self.post_process_function is None or self.compute_metrics is None: return output _lowercase : Dict = self.post_process_function(lowerCamelCase, lowerCamelCase, lowerCamelCase, 'predict') _lowercase : Optional[Any] = self.compute_metrics(lowerCamelCase) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(F'''{metric_key_prefix}_'''): _lowercase : Optional[Any] = metrics.pop(lowerCamelCase) metrics.update(output.metrics) return PredictionOutput(predictions=predictions.predictions, label_ids=predictions.label_ids, metrics=lowerCamelCase)
89
"""simple docstring""" from jiwer import compute_measures import datasets __lowerCAmelCase : Tuple = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __lowerCAmelCase : Union[str, Any] = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' __lowerCAmelCase : Optional[int] = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", ] , ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=False ) -> Optional[Any]: '''simple docstring''' if concatenate_texts: return compute_measures(_lowercase , _lowercase )["wer"] else: snake_case_ : List[str] = 0 snake_case_ : Optional[int] = 0 for prediction, reference in zip(_lowercase , _lowercase ): snake_case_ : Optional[Any] = compute_measures(_lowercase , _lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
58
0
'''simple docstring''' import os from typing import List, Optional, Union from ...tokenization_utils import PreTrainedTokenizer from ...tokenization_utils_base import AddedToken from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = {'''vocab_file''': '''vocab.txt'''} __UpperCAmelCase = { '''vocab_file''': { '''facebook/esm2_t6_8M_UR50D''': '''https://huggingface.co/facebook/esm2_t6_8M_UR50D/resolve/main/vocab.txt''', '''facebook/esm2_t12_35M_UR50D''': '''https://huggingface.co/facebook/esm2_t12_35M_UR50D/resolve/main/vocab.txt''', }, } __UpperCAmelCase = { '''facebook/esm2_t6_8M_UR50D''': 1_024, '''facebook/esm2_t12_35M_UR50D''': 1_024, } def _snake_case ( A ) -> Optional[Any]: with open(A , '''r''' ) as f: lowerCAmelCase__ = f.read().splitlines() return [l.strip() for l in lines] class a__ ( a__ ): '''simple docstring''' lowercase__ : Optional[Any] = VOCAB_FILES_NAMES lowercase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , lowerCamelCase_ , lowerCamelCase_="<unk>" , lowerCamelCase_="<cls>" , lowerCamelCase_="<pad>" , lowerCamelCase_="<mask>" , lowerCamelCase_="<eos>" , **lowerCamelCase_ , ) -> Tuple: super().__init__(**lowerCamelCase_ ) lowerCAmelCase__ = load_vocab_file(lowerCamelCase_ ) lowerCAmelCase__ = dict(enumerate(self.all_tokens ) ) lowerCAmelCase__ = {tok: ind for ind, tok in enumerate(self.all_tokens )} lowerCAmelCase__ = unk_token lowerCAmelCase__ = cls_token lowerCAmelCase__ = pad_token lowerCAmelCase__ = mask_token lowerCAmelCase__ = eos_token lowerCAmelCase__ = self.all_tokens self._create_trie(self.unique_no_split_tokens ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> str: return self._id_to_token.get(lowerCamelCase_ , self.unk_token ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> int: return self._token_to_id.get(lowerCamelCase_ , self._token_to_id.get(self.unk_token ) ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , **lowerCamelCase_ ) -> Union[str, Any]: return text.split() def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_=False ) -> Dict: return len(self._id_to_token ) def __SCREAMING_SNAKE_CASE ( self ) -> int: return {token: i for i, token in enumerate(self.all_tokens )} def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> int: return self._token_to_id.get(lowerCamelCase_ , self._token_to_id.get(self.unk_token ) ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> str: return self._id_to_token.get(lowerCamelCase_ , self.unk_token ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ = None ) -> List[int]: lowerCAmelCase__ = [self.cls_token_id] lowerCAmelCase__ = [self.eos_token_id] # No sep token in ESM vocabulary if token_ids_a is None: if self.eos_token_id is None: return cls + token_ids_a else: return cls + token_ids_a + sep elif self.eos_token_id is None: raise ValueError('''Cannot tokenize multiple sequences when EOS token is not set!''' ) return cls + token_ids_a + sep + token_ids_a + sep # Multiple inputs always have an EOS token def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False ) -> List[int]: if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if token in self.all_special_ids else 0 for token in token_ids_a] lowerCAmelCase__ = [1] + ([0] * len(lowerCamelCase_ )) + [1] if token_ids_a is not None: mask += [0] * len(lowerCamelCase_ ) + [1] return mask def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ ) -> Union[str, Any]: lowerCAmelCase__ = os.path.join(lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + '''vocab.txt''' ) with open(lowerCamelCase_ , '''w''' ) as f: f.write('''\n'''.join(self.all_tokens ) ) return (vocab_file,) @property def __SCREAMING_SNAKE_CASE ( self ) -> int: return self.get_vocab_size(with_added_tokens=lowerCamelCase_ ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ = False ) -> int: return super()._add_tokens(lowerCamelCase_ , special_tokens=lowerCamelCase_ )
90
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1_3 , _lowercase=3 , _lowercase=2_2_4 , _lowercase=3_0 , _lowercase=4_0_0 , _lowercase=True , _lowercase=None , _lowercase=True , _lowercase=[0.5, 0.5, 0.5] , _lowercase=[0.5, 0.5, 0.5] , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : Optional[Any] = batch_size snake_case_ : Dict = num_channels snake_case_ : Optional[Any] = image_size snake_case_ : Optional[Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[int] = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : Dict = image_std def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ViTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = EfficientFormerImageProcessorTester(self ) @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.image_proc_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , """image_mean""" ) ) self.assertTrue(hasattr(_lowercase , """image_std""" ) ) self.assertTrue(hasattr(_lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowercase , """do_resize""" ) ) self.assertTrue(hasattr(_lowercase , """size""" ) ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' pass def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input snake_case_ : Any = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Optional[Any] = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Dict = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input snake_case_ : int = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : int = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input snake_case_ : Dict = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Tuple = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
58
0
"""simple docstring""" from math import sqrt def _snake_case ( snake_case__ : int ): A = 0 for i in range(1 , int(sqrt(snake_case__ ) + 1 ) ): if n % i == 0 and i != sqrt(snake_case__ ): total += i + n // i elif i == sqrt(snake_case__ ): total += i return total - n def _snake_case ( snake_case__ : int = 1_0000 ): A = sum( i for i in range(1 , snake_case__ ) if sum_of_divisors(sum_of_divisors(snake_case__ ) ) == i and sum_of_divisors(snake_case__ ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
91
"""simple docstring""" from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCAmelCase : int = TypeVar('''KT''') __lowerCAmelCase : Union[str, Any] = TypeVar('''VT''') class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = "root" , _lowercase = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = key snake_case_ : Tuple = value snake_case_ : list[Node[KT, VT]] = [] def __repr__( self ) -> str: '''simple docstring''' return f'Node({self.key}: {self.value})' @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return len(self.forward ) class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = 0.5 , _lowercase = 1_6 ) -> int: '''simple docstring''' snake_case_ : Node[KT, VT] = Node[KT, VT]() snake_case_ : Union[str, Any] = 0 snake_case_ : Optional[int] = p snake_case_ : Any = max_level def __str__( self ) -> str: '''simple docstring''' snake_case_ : str = list(self ) if len(_lowercase ) == 0: return f'SkipList(level={self.level})' snake_case_ : List[Any] = max((len(str(_lowercase ) ) for item in items) , default=4 ) snake_case_ : str = max(_lowercase , 4 ) + 4 snake_case_ : Union[str, Any] = self.head snake_case_ : Dict = [] snake_case_ : List[str] = node.forward.copy() lines.append(f'[{node.key}]'.ljust(_lowercase , """-""" ) + """* """ * len(_lowercase ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) while len(node.forward ) != 0: snake_case_ : Optional[Any] = node.forward[0] lines.append( f'[{node.key}]'.ljust(_lowercase , """-""" ) + """ """.join(str(n.key ) if n.key == node.key else """|""" for n in forwards ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) snake_case_ : List[str] = node.forward lines.append("""None""".ljust(_lowercase ) + """* """ * len(_lowercase ) ) return f'SkipList(level={self.level})\n' + "\n".join(_lowercase ) def __iter__( self ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.head while len(node.forward ) != 0: yield node.forward[0].key snake_case_ : Dict = node.forward[0] def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[int] = 1 while random() < self.p and level < self.max_level: level += 1 return level def UpperCAmelCase__ ( self , _lowercase ) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: snake_case_ : List[Any] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(_lowercase ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: for i, update_node in enumerate(_lowercase ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: snake_case_ : List[str] = node.forward[i] else: snake_case_ : Tuple = update_node.forward[:i] def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> str: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: snake_case_ : List[Any] = value else: snake_case_ : Optional[int] = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , _lowercase ): update_vector.append(self.head ) snake_case_ : Any = level snake_case_ : Optional[int] = Node(_lowercase , _lowercase ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(_lowercase ) else: snake_case_ : Optional[Any] = new_node def UpperCAmelCase__ ( self , _lowercase ) -> VT | None: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: return node.value return None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : List[str] = SkipList() skip_list.insert("""Key1""" , 3 ) skip_list.insert("""Key2""" , 1_2 ) skip_list.insert("""Key3""" , 4_1 ) skip_list.insert("""Key4""" , -1_9 ) snake_case_ : Optional[int] = skip_list.head snake_case_ : List[Any] = {} while node.level != 0: snake_case_ : List[str] = node.forward[0] snake_case_ : Union[str, Any] = node.value assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 1_2 assert all_values["Key3"] == 4_1 assert all_values["Key4"] == -1_9 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_0 ) skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""Key5""" , 7 ) skip_list.insert("""Key7""" , 1_0 ) skip_list.insert("""Key10""" , 5 ) skip_list.insert("""Key7""" , 7 ) skip_list.insert("""Key5""" , 5 ) skip_list.insert("""Key10""" , 1_0 ) snake_case_ : str = skip_list.head snake_case_ : str = {} while node.level != 0: snake_case_ : Optional[Any] = node.forward[0] snake_case_ : int = node.value if len(__UpperCamelCase ) != 4: print() assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 1_2 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 1_0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : str = SkipList() assert skip_list.find("""Some key""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[Any] = SkipList() skip_list.insert("""Key2""" , 2_0 ) assert skip_list.find("""Key2""" ) == 2_0 skip_list.insert("""Some Key""" , 1_0 ) skip_list.insert("""Key2""" , 8 ) skip_list.insert("""V""" , 1_3 ) assert skip_list.find("""Y""" ) is None assert skip_list.find("""Key2""" ) == 8 assert skip_list.find("""Some Key""" ) == 1_0 assert skip_list.find("""V""" ) == 1_3 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Any = SkipList() skip_list.delete("""Some key""" ) assert len(skip_list.head.forward ) == 0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Tuple = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) == 1_4 assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""X""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key1""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Union[str, Any] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4_2 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""X""" ) def traverse_keys(__UpperCamelCase : str ): yield node.key for forward_node in node.forward: yield from traverse_keys(__UpperCamelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __lowerCAmelCase ( ): '''simple docstring''' def is_sorted(__UpperCamelCase : List[Any] ): return all(next_item >= item for item, next_item in zip(__UpperCamelCase , lst[1:] ) ) snake_case_ : str = SkipList() for i in range(1_0 ): skip_list.insert(__UpperCamelCase , __UpperCamelCase ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.insert(-1_2 , -1_2 ) skip_list.insert(7_7 , 7_7 ) assert is_sorted(list(__UpperCamelCase ) ) def __lowerCAmelCase ( ): '''simple docstring''' for _ in range(1_0_0 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = SkipList() skip_list.insert(2 , """2""" ) skip_list.insert(4 , """4""" ) skip_list.insert(6 , """4""" ) skip_list.insert(4 , """5""" ) skip_list.insert(8 , """4""" ) skip_list.insert(9 , """4""" ) skip_list.delete(4 ) print(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
58
0
'''simple docstring''' from string import ascii_lowercase, ascii_uppercase def _lowerCAmelCase ( __magic_name__ : str ) -> str: if not sentence: return "" lowercase : Tuple =dict(zip(__magic_name__ , __magic_name__ ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __name__ == "__main__": from doctest import testmod testmod()
92
"""simple docstring""" import argparse import os import re import packaging.version __lowerCAmelCase : Optional[Any] = '''examples/''' __lowerCAmelCase : Union[str, Any] = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __lowerCAmelCase : Union[str, Any] = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } __lowerCAmelCase : List[Any] = '''README.md''' def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : Dict ): '''simple docstring''' with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : Any = f.read() snake_case_ , snake_case_ : Optional[int] = REPLACE_PATTERNS[pattern] snake_case_ : Union[str, Any] = replace.replace("""VERSION""" , __UpperCamelCase ) snake_case_ : List[Any] = re_pattern.sub(__UpperCamelCase , __UpperCamelCase ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' for folder, directories, fnames in os.walk(__UpperCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(__UpperCamelCase , __UpperCamelCase ) , __UpperCamelCase , pattern="""examples""" ) def __lowerCAmelCase ( __UpperCamelCase : List[str] , __UpperCamelCase : int=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not patch: update_version_in_examples(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = """🤗 Transformers currently provides the following architectures""" snake_case_ : Union[str, Any] = """1. Want to contribute a new model?""" with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : str = f.readlines() # Find the start of the list. snake_case_ : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 snake_case_ : Optional[int] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): snake_case_ : Any = lines[index].replace( """https://huggingface.co/docs/diffusers/main/model_doc""" , """https://huggingface.co/docs/diffusers/model_doc""" , ) index += 1 with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] , """r""" ) as f: snake_case_ : Any = f.read() snake_case_ : Tuple = REPLACE_PATTERNS["""init"""][0].search(__UpperCamelCase ).groups()[0] return packaging.version.parse(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : str=False ): '''simple docstring''' snake_case_ : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: snake_case_ : str = default_version.base_version elif patch: snake_case_ : str = F'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: snake_case_ : str = F'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. snake_case_ : int = input(F'Which version are you releasing? [{default_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Optional[int] = default_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase , patch=__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = get_version() snake_case_ : str = F'{current_version.major}.{current_version.minor + 1}.0.dev0' snake_case_ : Tuple = current_version.base_version # Check with the user we got that right. snake_case_ : Optional[int] = input(F'Which version are we developing now? [{dev_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Dict = dev_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": __lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __lowerCAmelCase : str = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
58
0
"""simple docstring""" import os import re import shutil import sys import tempfile import unittest import black __A = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If DDPMSchedulerOutput is changed in scheduling_ddpm.py, this code needs to be manually updated. __A = """ \"\"\" Output class for the scheduler's step function output. Args: prev_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images): Computed sample (x_{t-1}) of previous timestep. `prev_sample` should be used as next model input in the denoising loop. pred_original_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images): The predicted denoised sample (x_{0}) based on the model output from the current timestep. `pred_original_sample` can be used to preview progress or for guidance. \"\"\" prev_sample: torch.FloatTensor pred_original_sample: Optional[torch.FloatTensor] = None """ class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = tempfile.mkdtemp() os.makedirs(os.path.join(self.diffusers_dir , 'schedulers/' ) ) lowerCAmelCase__ :Optional[int] = self.diffusers_dir shutil.copy( os.path.join(__UpperCAmelCase , 'src/diffusers/schedulers/scheduling_ddpm.py' ) , os.path.join(self.diffusers_dir , 'schedulers/scheduling_ddpm.py' ) , ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = 'src/diffusers' shutil.rmtree(self.diffusers_dir ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' lowerCAmelCase__ :Any = comment + F"\nclass {class_name}(nn.Module):\n" + class_code if overwrite_result is not None: lowerCAmelCase__ :str = comment + F"\nclass {class_name}(nn.Module):\n" + overwrite_result lowerCAmelCase__ :List[Any] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_1_9 ) lowerCAmelCase__ :str = black.format_str(__UpperCAmelCase , mode=__UpperCAmelCase ) lowerCAmelCase__ :List[Any] = os.path.join(self.diffusers_dir , 'new_code.py' ) with open(__UpperCAmelCase , 'w' , newline='\n' ) as f: f.write(__UpperCAmelCase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(__UpperCAmelCase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=__UpperCAmelCase ) with open(__UpperCAmelCase , 'r' ) as f: self.assertTrue(f.read() , __UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = check_copies.find_code_in_diffusers('schedulers.scheduling_ddpm.DDPMSchedulerOutput' ) self.assertEqual(__UpperCAmelCase , __UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' self.check_copy_consistency( '# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput' , 'DDPMSchedulerOutput' , REFERENCE_CODE + '\n' , ) # With no empty line at the end self.check_copy_consistency( '# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput' , 'DDPMSchedulerOutput' , __UpperCAmelCase , ) # Copy consistency with rename self.check_copy_consistency( '# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test' , 'TestSchedulerOutput' , re.sub('DDPM' , 'Test' , __UpperCAmelCase ) , ) # Copy consistency with a really long name lowerCAmelCase__ :List[str] = 'TestClassWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason' self.check_copy_consistency( F"# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->{long_class_name}" , F"{long_class_name}SchedulerOutput" , re.sub('Bert' , __UpperCAmelCase , __UpperCAmelCase ) , ) # Copy consistency with overwrite self.check_copy_consistency( '# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test' , 'TestSchedulerOutput' , __UpperCAmelCase , overwrite_result=re.sub('DDPM' , 'Test' , __UpperCAmelCase ) , )
93
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
58
0
'''simple docstring''' import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class UpperCAmelCase_ ( __A , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = ReformerTokenizer UpperCamelCase_ = ReformerTokenizerFast UpperCamelCase_ = True UpperCamelCase_ = False UpperCamelCase_ = True def A__ ( self : Dict ) -> Tuple: '''simple docstring''' super().setUp() lowercase : Optional[Any] =ReformerTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def A__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' lowercase : List[Any] ='''<s>''' lowercase : List[Any] =1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase ) , UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase ) , UpperCAmelCase ) def A__ ( self : Optional[Any] ) -> Any: '''simple docstring''' lowercase : Any =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(UpperCAmelCase ) , 1000 ) def A__ ( self : Tuple ) -> str: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def A__ ( self : List[str] ) -> Tuple: '''simple docstring''' if not self.test_rust_tokenizer: return lowercase : List[str] =self.get_tokenizer() lowercase : int =self.get_rust_tokenizer() lowercase : Any ='''I was born in 92000, and this is falsé.''' lowercase : Optional[Any] =tokenizer.tokenize(UpperCAmelCase ) lowercase : List[Any] =rust_tokenizer.tokenize(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) lowercase : List[Any] =tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) lowercase : Optional[int] =rust_tokenizer.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) lowercase : Optional[Any] =self.get_rust_tokenizer() lowercase : str =tokenizer.encode(UpperCAmelCase ) lowercase : Any =rust_tokenizer.encode(UpperCAmelCase ) self.assertListEqual(UpperCAmelCase , UpperCAmelCase ) def A__ ( self : Tuple , UpperCAmelCase : int=15 ) -> Optional[int]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowercase : Tuple =self.rust_tokenizer_class.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) # Simple input lowercase : List[str] ='''This is a simple input''' lowercase : List[Any] =['''This is a simple input 1''', '''This is a simple input 2'''] lowercase : Optional[int] =('''This is a simple input''', '''This is a pair''') lowercase : List[Any] =[ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(UpperCAmelCase , tokenizer_r.encode , UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' ) # Simple input self.assertRaises(UpperCAmelCase , tokenizer_r.encode_plus , UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' ) # Simple input self.assertRaises( UpperCAmelCase , tokenizer_r.batch_encode_plus , UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' , ) # Pair input self.assertRaises(UpperCAmelCase , tokenizer_r.encode , UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' ) # Pair input self.assertRaises(UpperCAmelCase , tokenizer_r.encode_plus , UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' ) # Pair input self.assertRaises( UpperCAmelCase , tokenizer_r.batch_encode_plus , UpperCAmelCase , max_length=UpperCAmelCase , padding='''max_length''' , ) def A__ ( self : str ) -> int: '''simple docstring''' pass def A__ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowercase : Any =ReformerTokenizer(UpperCAmelCase , keep_accents=UpperCAmelCase ) lowercase : str =tokenizer.tokenize('''This is a test''' ) self.assertListEqual(UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase ) , [285, 46, 10, 170, 382] , ) lowercase : Any =tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) lowercase : Dict =tokenizer.convert_tokens_to_ids(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowercase : str =tokenizer.convert_ids_to_tokens(UpperCAmelCase ) self.assertListEqual( UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def A__ ( self : str ) -> Optional[int]: '''simple docstring''' return ReformerTokenizer.from_pretrained('''google/reformer-crime-and-punishment''' ) @slow def A__ ( self : str ) -> List[Any]: '''simple docstring''' lowercase : Tuple ='''Hello World!''' lowercase : Any =[126, 32, 262, 152, 38, 72, 287] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @slow def A__ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' lowercase : str =( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) lowercase : Union[str, Any] =[ 108, 265, 24, 111, 4, 258, 156, 35, 28, 275, 3, 259, 297, 260, 84, 4, 35, 110, 44, 8, 259, 91, 268, 21, 11, 209, 274, 109, 266, 277, 117, 86, 93, 315, 258, 278, 258, 277, 258, 0, 258, 288, 258, 319, 258, 0, 258, 0, 258, 0, 258, 0, 258, 287, 258, 315, 258, 289, 258, 278, 99, 269, 266, 262, 8, 259, 241, 4, 217, 230, 268, 266, 55, 168, 106, 75, 193, 266, 223, 27, 49, 26, 282, 25, 264, 299, 19, 26, 0, 258, 277, 117, 86, 93, 176, 183, 270, 11, 262, 42, 61, 265, ] self.assertListEqual(UpperCAmelCase , self.big_tokenizer.encode(UpperCAmelCase ) ) @require_torch @slow def A__ ( self : Optional[Any] ) -> List[Any]: '''simple docstring''' import torch from transformers import ReformerConfig, ReformerModel # Build sequence lowercase : List[Any] =list(self.big_tokenizer.get_vocab().keys() )[:10] lowercase : Union[str, Any] =''' '''.join(UpperCAmelCase ) lowercase : int =self.big_tokenizer.encode_plus(UpperCAmelCase , return_tensors='''pt''' ) lowercase : List[str] =self.big_tokenizer.batch_encode_plus([sequence, sequence] , return_tensors='''pt''' ) lowercase : Optional[Any] =ReformerConfig() # The input gets padded during training so adjust the axial position encodings from the pretrained model value of (512, 1024) lowercase : str =encoded_sequence['''input_ids'''].shape lowercase : Any =ReformerModel(UpperCAmelCase ) # Reformer has config.vocab_size == tokenizer.vocab_size == len(tokenizer) - 1 = 320; len(tokenizer) is 321 (including a pad token with id 320) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**UpperCAmelCase ) model(**UpperCAmelCase ) @slow def A__ ( self : int ) -> Dict: '''simple docstring''' lowercase : Optional[Any] ={'''input_ids''': [[108, 265, 24, 111, 4, 258, 156, 7, 51, 279, 58, 7, 76, 25, 69, 278], [140, 243, 264, 134, 17, 267, 77, 263, 22, 262, 297, 258, 304, 177, 279, 266, 14, 89, 13, 35, 261, 299, 272, 137, 275, 278]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # This tokenizer does not know some characters like ")". # That is the reason why we use very simple texts here. # Also see https://github.com/huggingface/transformers/pull/11737#issuecomment-850769064 lowercase : Optional[Any] =[ '''This is a very simple sentence.''', '''The quick brown fox jumps over the lazy dog.''', ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase , model_name='''google/reformer-crime-and-punishment''' , revision='''0e6c3decb8211d49bf881013425dc8b0448b3f5a''' , padding=UpperCAmelCase , sequences=UpperCAmelCase , )
94
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) snake_case_ : str = precision snake_case_ : Any = ceil(precision / 1_4 ) snake_case_ : Dict = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : Optional[Any] = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : Optional[int] = Decimal(__UpperCamelCase ) for k in range(1 , __UpperCamelCase ): snake_case_ : Any = factorial(6 * k ) // (factorial(3 * k ) * factorial(__UpperCamelCase ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __lowerCAmelCase : int = 50 print(F'''The first {n} digits of pi is: {pi(n)}''')
58
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = torch.device('''cpu''') def snake_case ( ): UpperCAmelCase_ : str = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(A__ ,stream=A__ ).raw ) return im def snake_case ( A__ ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1703e00, 2.1107e00, -2.0811e00, 8.8685e-01, 2.4360e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9636e-01, 2.3478e-01, -1.6963e00, -1.7381e00, -8.6337e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2768e-01, -4.7429e-01, -1.0897e00, -1.0248e00, 3.5523e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5330e-01, 2.4211e-01, -6.0185e-01, -8.2789e-01, -6.0446e-02] ) def snake_case ( A__ ,A__ ,A__ ): UpperCAmelCase_ : Tuple = dct.pop(A__ ) UpperCAmelCase_ : Optional[Any] = val def snake_case ( A__ ): UpperCAmelCase_ : List[str] = [] for k in state_dict.keys(): UpperCAmelCase_ : Union[str, Any] = k if ".pwconv" in k: UpperCAmelCase_ : Dict = k_new.replace(".pwconv" ,".point_wise_conv" ) if ".dwconv" in k: UpperCAmelCase_ : Any = k_new.replace(".dwconv" ,".depth_wise_conv" ) if ".Proj." in k: UpperCAmelCase_ : Dict = k_new.replace(".Proj." ,".proj." ) if "patch_embed" in k_new: UpperCAmelCase_ : Tuple = k_new.replace("patch_embed" ,"swiftformer.patch_embed.patch_embedding" ) if "network" in k_new: UpperCAmelCase_ : List[Any] = k_new.split("." ) if ls[2].isdigit(): UpperCAmelCase_ : Tuple = "swiftformer.encoder.network." + ls[1] + ".blocks." + ls[2] + "." + ".".join(ls[3:] ) else: UpperCAmelCase_ : Optional[Any] = k_new.replace("network" ,"swiftformer.encoder.network" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def snake_case ( A__ ,A__ ,A__ ): UpperCAmelCase_ : Optional[int] = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Optional[Any] = 10_00 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[str] = json.load(open(hf_hub_download(A__ ,A__ ,repo_type="dataset" ) ,"r" ) ) UpperCAmelCase_ : Tuple = {int(A__ ): v for k, v in idalabel.items()} UpperCAmelCase_ : List[Any] = idalabel UpperCAmelCase_ : Optional[Any] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": UpperCAmelCase_ : Tuple = [3, 3, 6, 4] UpperCAmelCase_ : str = [48, 56, 1_12, 2_20] elif swiftformer_name == "swiftformer_s": UpperCAmelCase_ : Optional[Any] = [3, 3, 9, 6] UpperCAmelCase_ : Optional[Any] = [48, 64, 1_68, 2_24] elif swiftformer_name == "swiftformer_l1": UpperCAmelCase_ : int = [4, 3, 10, 5] UpperCAmelCase_ : Union[str, Any] = [48, 96, 1_92, 3_84] elif swiftformer_name == "swiftformer_l3": UpperCAmelCase_ : Dict = [4, 4, 12, 6] UpperCAmelCase_ : Optional[int] = [64, 1_28, 3_20, 5_12] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("https" ): UpperCAmelCase_ : List[Any] = torch.hub.load_state_dict_from_url(A__ ,map_location="cpu" ,check_hash=A__ ) else: UpperCAmelCase_ : Any = torch.load(A__ ,map_location="cpu" ) UpperCAmelCase_ : List[str] = checkpoint UpperCAmelCase_ : Dict = create_rename_keys(A__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(A__ ,A__ ,A__ ) # load HuggingFace model UpperCAmelCase_ : Optional[int] = SwiftFormerForImageClassification(A__ ).eval() hf_model.load_state_dict(A__ ) # prepare test inputs UpperCAmelCase_ : Tuple = prepare_img() UpperCAmelCase_ : int = ViTImageProcessor.from_pretrained("preprocessor_config" ) UpperCAmelCase_ : int = processor(images=A__ ,return_tensors="pt" ) # compare outputs from both models UpperCAmelCase_ : List[Any] = get_expected_output(A__ ) UpperCAmelCase_ : int = hf_model(inputs["pixel_values"] ).logits assert hf_logits.shape == torch.Size([1, 10_00] ) assert torch.allclose(hf_logits[0, 0:5] ,A__ ,atol=1e-3 ) Path(A__ ).mkdir(exist_ok=A__ ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(A__ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swiftformer_name''', default='''swiftformer_xs''', choices=['''swiftformer_xs''', '''swiftformer_s''', '''swiftformer_l1''', '''swiftformer_l3'''], type=str, help='''Name of the SwiftFormer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''./converted_outputs/''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--original_ckpt''', default=None, type=str, help='''Path to the original model checkpoint.''') lowerCamelCase_ = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
95
"""simple docstring""" import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCAmelCase ( __UpperCamelCase : Optional[int] ): '''simple docstring''' snake_case_ : Any = torch.exp(__UpperCamelCase ) snake_case_ : Optional[int] = torch.sum(__UpperCamelCase , dim=1 ) # sum of exp(x_i) snake_case_ : str = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(__UpperCamelCase ) - B / A class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> int: '''simple docstring''' super().__init__() snake_case_ : Tuple = config.output_attentions snake_case_ : str = config.output_hidden_states snake_case_ : List[str] = nn.ModuleList([BertLayer(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Tuple = nn.ModuleList([BertHighway(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Any = [-1 for _ in range(config.num_hidden_layers )] def UpperCAmelCase__ ( self , _lowercase ) -> Tuple: '''simple docstring''' if (type(_lowercase ) is float) or (type(_lowercase ) is int): for i in range(len(self.early_exit_entropy ) ): snake_case_ : Dict = x else: snake_case_ : Union[str, Any] = x def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : int = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Any: '''simple docstring''' snake_case_ : str = () snake_case_ : str = () snake_case_ : List[str] = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: snake_case_ : int = all_hidden_states + (hidden_states,) snake_case_ : Any = layer_module( _lowercase , _lowercase , head_mask[i] , _lowercase , _lowercase ) snake_case_ : Dict = layer_outputs[0] if self.output_attentions: snake_case_ : str = all_attentions + (layer_outputs[1],) snake_case_ : Optional[int] = (hidden_states,) if self.output_hidden_states: snake_case_ : Tuple = current_outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : int = current_outputs + (all_attentions,) snake_case_ : Optional[Any] = self.highway[i](_lowercase ) # logits, pooled_output if not self.training: snake_case_ : Tuple = highway_exit[0] snake_case_ : List[str] = entropy(_lowercase ) snake_case_ : str = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy snake_case_ : Union[str, Any] = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: snake_case_ : List[Any] = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(_lowercase , i + 1 ) else: snake_case_ : Dict = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: snake_case_ : Dict = all_hidden_states + (hidden_states,) snake_case_ : str = (hidden_states,) if self.output_hidden_states: snake_case_ : List[Any] = outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : Union[str, Any] = outputs + (all_attentions,) snake_case_ : List[str] = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( '''The Bert Model transformer with early exiting (DeeBERT). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> Union[str, Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config snake_case_ : int = BertEmbeddings(_lowercase ) snake_case_ : Tuple = DeeBertEncoder(_lowercase ) snake_case_ : int = BertPooler(_lowercase ) self.init_weights() def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' self.encoder.init_highway_pooler(self.pooler ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return self.embeddings.word_embeddings def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = value def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(_lowercase ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' if input_ids is not None and inputs_embeds is not None: raise ValueError("""You cannot specify both input_ids and inputs_embeds at the same time""" ) elif input_ids is not None: snake_case_ : Dict = input_ids.size() elif inputs_embeds is not None: snake_case_ : Optional[int] = inputs_embeds.size()[:-1] else: raise ValueError("""You have to specify either input_ids or inputs_embeds""" ) snake_case_ : int = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: snake_case_ : Dict = torch.ones(_lowercase , device=_lowercase ) if encoder_attention_mask is None: snake_case_ : Tuple = torch.ones(_lowercase , device=_lowercase ) if token_type_ids is None: snake_case_ : Any = torch.zeros(_lowercase , dtype=torch.long , device=_lowercase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. snake_case_ : torch.Tensor = self.get_extended_attention_mask(_lowercase , _lowercase , _lowercase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: snake_case_ : List[str] = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: snake_case_ : Any = encoder_attention_mask[:, None, None, :] snake_case_ : List[str] = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility snake_case_ : List[str] = (1.0 - encoder_extended_attention_mask) * -1_0000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] snake_case_ : int = self.get_head_mask(_lowercase , self.config.num_hidden_layers ) snake_case_ : List[str] = self.embeddings( input_ids=_lowercase , position_ids=_lowercase , token_type_ids=_lowercase , inputs_embeds=_lowercase ) snake_case_ : List[str] = self.encoder( _lowercase , attention_mask=_lowercase , head_mask=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , ) snake_case_ : Optional[Any] = encoder_outputs[0] snake_case_ : Union[str, Any] = self.pooler(_lowercase ) snake_case_ : Optional[Any] = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = message snake_case_ : str = exit_layer # start from 1! class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__() snake_case_ : str = BertPooler(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Dict = nn.Linear(config.hidden_size , config.num_labels ) def UpperCAmelCase__ ( self , _lowercase ) -> Dict: '''simple docstring''' snake_case_ : List[str] = encoder_outputs[0] snake_case_ : List[Any] = self.pooler(_lowercase ) # "return" pooler_output # BertModel snake_case_ : Tuple = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification snake_case_ : Union[str, Any] = bmodel_output[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : List[str] = self.classifier(_lowercase ) return logits, pooled_output @add_start_docstrings( '''Bert Model (with early exiting - DeeBERT) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config.num_labels snake_case_ : Tuple = config.num_hidden_layers snake_case_ : Any = DeeBertModel(_lowercase ) snake_case_ : Optional[int] = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Tuple = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> int: '''simple docstring''' snake_case_ : int = self.num_layers try: snake_case_ : Any = self.bert( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits snake_case_ : str = outputs[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : List[Any] = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : Optional[int] = e.message snake_case_ : Dict = e.exit_layer snake_case_ : Optional[Any] = outputs[0] if not self.training: snake_case_ : int = entropy(_lowercase ) snake_case_ : int = [] snake_case_ : List[str] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : Dict = [] for highway_exit in outputs[-1]: snake_case_ : List[Any] = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : List[Any] = MSELoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : List[str] = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : str = (loss,) + outputs if not self.training: snake_case_ : Optional[int] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : str = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
58
0
"""simple docstring""" from __future__ import annotations def a ( __UpperCAmelCase : list[float] ) -> bool: if len(__UpperCAmelCase ) < 2: raise ValueError("""Monogons and Digons are not polygons in the Euclidean space""" ) if any(i <= 0 for i in nums ): raise ValueError("""All values must be greater than 0""" ) __magic_name__: List[Any] = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' return 1 / (1 + np.exp(-z )) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int ): '''simple docstring''' return (-y * np.log(__UpperCamelCase ) - (1 - y) * np.log(1 - h )).mean() def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : List[Any] , __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Optional[int] = np.dot(__UpperCamelCase , __UpperCamelCase ) return np.sum(y * scores - np.log(1 + np.exp(__UpperCamelCase ) ) ) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : int=7_0_0_0_0 ): '''simple docstring''' snake_case_ : Dict = np.zeros(x.shape[1] ) for iterations in range(__UpperCamelCase ): snake_case_ : Any = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Optional[Any] = np.dot(x.T , h - y ) / y.size snake_case_ : str = theta - alpha * gradient # updating the weights snake_case_ : int = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Dict = cost_function(__UpperCamelCase , __UpperCamelCase ) if iterations % 1_0_0 == 0: print(F'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": __lowerCAmelCase : Any = datasets.load_iris() __lowerCAmelCase : List[Any] = iris.data[:, :2] __lowerCAmelCase : Tuple = (iris.target != 0) * 1 __lowerCAmelCase : Any = 0.1 __lowerCAmelCase : List[Any] = logistic_reg(alpha, x, y, max_iterations=7_0000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' return sigmoid_function( np.dot(__UpperCamelCase , __UpperCamelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = (x[:, 0].min(), x[:, 0].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Tuple = (x[:, 1].min(), x[:, 1].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Optional[Any] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) __lowerCAmelCase : Any = np.c_[xxa.ravel(), xxa.ravel()] __lowerCAmelCase : Optional[int] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
58
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def a ( snake_case__: Dict ): '''simple docstring''' lowercase_ = [2, 2, 6, 2] if '''tiny''' in model_name else [2, 2, 18, 2] lowercase_ = True if '''large''' in model_name or '''huge''' in model_name else False lowercase_ = True if '''large''' in model_name or '''huge''' in model_name else False lowercase_ = True if '''large''' in model_name or '''huge''' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: lowercase_ = [3, 3, 3, 3] lowercase_ = [5, 5, 5, 5] elif "fl4" in model_name: lowercase_ = [4, 4, 4, 4] lowercase_ = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: lowercase_ = [3, 3, 3, 3] if "lrf" in model_name: lowercase_ = [3, 3, 3, 3] else: lowercase_ = [2, 2, 2, 2] if "tiny" in model_name: lowercase_ = 96 elif "small" in model_name: lowercase_ = 96 elif "base" in model_name: lowercase_ = 128 elif "large" in model_name: lowercase_ = 192 elif "xlarge" in model_name: lowercase_ = 256 elif "huge" in model_name: lowercase_ = 352 # set label information lowercase_ = '''huggingface/label-files''' if "large" in model_name or "huge" in model_name: lowercase_ = '''imagenet-22k-id2label.json''' else: lowercase_ = '''imagenet-1k-id2label.json''' lowercase_ = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type='''dataset''' ) , '''r''' ) ) lowercase_ = {int(snake_case__ ): v for k, v in idalabel.items()} lowercase_ = {v: k for k, v in idalabel.items()} lowercase_ = FocalNetConfig( embed_dim=snake_case__ , depths=snake_case__ , focal_levels=snake_case__ , focal_windows=snake_case__ , use_conv_embed=snake_case__ , idalabel=snake_case__ , labelaid=snake_case__ , use_post_layernorm=snake_case__ , use_layerscale=snake_case__ , ) return config def a ( snake_case__: Any ): '''simple docstring''' if "patch_embed.proj" in name: lowercase_ = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: lowercase_ = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: lowercase_ = '''encoder.''' + name if "encoder.layers" in name: lowercase_ = name.replace('''encoder.layers''' , '''encoder.stages''' ) if "downsample.proj" in name: lowercase_ = name.replace('''downsample.proj''' , '''downsample.projection''' ) if "blocks" in name: lowercase_ = name.replace('''blocks''' , '''layers''' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: lowercase_ = name.replace('''modulation.f''' , '''modulation.projection_in''' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: lowercase_ = name.replace('''modulation.h''' , '''modulation.projection_context''' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: lowercase_ = name.replace('''modulation.proj''' , '''modulation.projection_out''' ) if name == "norm.weight": lowercase_ = '''layernorm.weight''' if name == "norm.bias": lowercase_ = '''layernorm.bias''' if "head" in name: lowercase_ = name.replace('''head''' , '''classifier''' ) else: lowercase_ = '''focalnet.''' + name return name def a ( snake_case__: str , snake_case__: str , snake_case__: Any=False ): '''simple docstring''' # fmt: off lowercase_ = { '''focalnet-tiny''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth''', '''focalnet-tiny-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth''', '''focalnet-small''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth''', '''focalnet-small-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth''', '''focalnet-base''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth''', '''focalnet-base-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth''', '''focalnet-large-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth''', '''focalnet-large-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth''', '''focalnet-xlarge-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth''', '''focalnet-xlarge-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth''', } # fmt: on lowercase_ = model_name_to_url[model_name] print('''Checkpoint URL: ''' , snake_case__ ) lowercase_ = torch.hub.load_state_dict_from_url(snake_case__ , map_location='''cpu''' )['''model'''] # rename keys for key in state_dict.copy().keys(): lowercase_ = state_dict.pop(snake_case__ ) lowercase_ = val lowercase_ = get_focalnet_config(snake_case__ ) lowercase_ = FocalNetForImageClassification(snake_case__ ) model.eval() # load state dict model.load_state_dict(snake_case__ ) # verify conversion lowercase_ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowercase_ = BitImageProcessor( do_resize=snake_case__ , size={'''shortest_edge''': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=snake_case__ , crop_size=224 , do_normalize=snake_case__ , image_mean=snake_case__ , image_std=snake_case__ , ) lowercase_ = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ) lowercase_ = processor(images=snake_case__ , return_tensors='''pt''' ) lowercase_ = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_8_5, 0.4_5_6, 0.4_0_6] , std=[0.2_2_9, 0.2_2_4, 0.2_2_5] ), ] ) lowercase_ = image_transforms(snake_case__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , snake_case__ , atol=1e-4 ) lowercase_ = model(**snake_case__ ) lowercase_ = outputs.logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) print('''First values of logits:''' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": lowercase_ = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ) elif model_name == "focalnet-tiny-lrf": lowercase_ = torch.tensor([1.1_6_6_9, 0.0_1_2_5, -0.1_6_9_5] ) elif model_name == "focalnet-small": lowercase_ = torch.tensor([0.4_9_1_7, -0.0_4_3_0, 0.1_3_4_1] ) elif model_name == "focalnet-small-lrf": lowercase_ = torch.tensor([-0.2_5_8_8, -0.5_3_4_2, -0.2_3_3_1] ) elif model_name == "focalnet-base": lowercase_ = torch.tensor([-0.1_6_5_5, -0.4_0_9_0, -0.1_7_3_0] ) elif model_name == "focalnet-base-lrf": lowercase_ = torch.tensor([0.5_3_0_6, -0.0_4_8_3, -0.3_9_2_8] ) assert torch.allclose(outputs.logits[0, :3] , snake_case__ , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and processor of {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) processor.save_pretrained(snake_case__ ) if push_to_hub: print(F'''Pushing model and processor of {model_name} to the hub...''' ) model.push_to_hub(F'''{model_name}''' ) processor.push_to_hub(F'''{model_name}''' ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='focalnet-tiny', type=str, help='Name of the FocalNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub.', ) __a = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
97
"""simple docstring""" import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput __lowerCAmelCase : Tuple = '''scheduler_config.json''' class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 1 _lowerCamelCase = 2 _lowerCamelCase = 3 _lowerCamelCase = 4 _lowerCamelCase = 5 @dataclass class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 42 class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = SCHEDULER_CONFIG_NAME _lowerCamelCase = ['''dtype'''] _lowerCamelCase = [] _lowerCamelCase = True @classmethod def UpperCAmelCase__ ( cls , _lowercase = None , _lowercase = None , _lowercase=False , **_lowercase , ) -> Any: '''simple docstring''' snake_case_ , snake_case_ : int = cls.load_config( pretrained_model_name_or_path=_lowercase , subfolder=_lowercase , return_unused_kwargs=_lowercase , **_lowercase , ) snake_case_ , snake_case_ : Dict = cls.from_config(_lowercase , return_unused_kwargs=_lowercase , **_lowercase ) if hasattr(_lowercase , """create_state""" ) and getattr(_lowercase , """has_state""" , _lowercase ): snake_case_ : Any = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCAmelCase__ ( self , _lowercase , _lowercase = False , **_lowercase ) -> Optional[Any]: '''simple docstring''' self.save_config(save_directory=_lowercase , push_to_hub=_lowercase , **_lowercase ) @property def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' return self._get_compatibles() @classmethod def UpperCAmelCase__ ( cls ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = list(set([cls.__name__] + cls._compatibles ) ) snake_case_ : str = importlib.import_module(__name__.split(""".""" )[0] ) snake_case_ : Optional[int] = [ getattr(_lowercase , _lowercase ) for c in compatible_classes_str if hasattr(_lowercase , _lowercase ) ] return compatible_classes def __lowerCAmelCase ( __UpperCamelCase : jnp.ndarray , __UpperCamelCase : Tuple[int] ): '''simple docstring''' assert len(__UpperCamelCase ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__UpperCamelCase ) - x.ndim) ) , __UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Any=0.999 , __UpperCamelCase : Optional[int]=jnp.floataa ): '''simple docstring''' def alpha_bar(__UpperCamelCase : Optional[int] ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 snake_case_ : Optional[Any] = [] for i in range(__UpperCamelCase ): snake_case_ : Dict = i / num_diffusion_timesteps snake_case_ : Union[str, Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__UpperCamelCase ) / alpha_bar(__UpperCamelCase ) , __UpperCamelCase ) ) return jnp.array(__UpperCamelCase , dtype=__UpperCamelCase ) @flax.struct.dataclass class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 @classmethod def UpperCAmelCase__ ( cls , _lowercase ) -> int: '''simple docstring''' snake_case_ : Any = scheduler.config if config.trained_betas is not None: snake_case_ : Optional[Any] = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": snake_case_ : int = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case_ : str = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case_ : int = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( f'beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}' ) snake_case_ : Optional[Any] = 1.0 - betas snake_case_ : Any = jnp.cumprod(_lowercase , axis=0 ) return cls( alphas=_lowercase , betas=_lowercase , alphas_cumprod=_lowercase , ) def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ : Tuple = state.alphas_cumprod snake_case_ : Optional[int] = alphas_cumprod[timesteps] ** 0.5 snake_case_ : Dict = sqrt_alpha_prod.flatten() snake_case_ : int = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) snake_case_ : Optional[Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 snake_case_ : Dict = sqrt_one_minus_alpha_prod.flatten() snake_case_ : Tuple = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : str = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : List[Any] = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
58
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : Union[str, Any] = logging.get_logger(__name__) lowercase__ : Dict = { 'facebook/vit-mae-base': 'https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : Optional[Any] = 'vit_mae' def __init__( self : str , lowerCAmelCase__ : Optional[int]=768 , lowerCAmelCase__ : Optional[int]=12 , lowerCAmelCase__ : List[Any]=12 , lowerCAmelCase__ : Dict=3072 , lowerCAmelCase__ : Optional[int]="gelu" , lowerCAmelCase__ : Tuple=0.0 , lowerCAmelCase__ : List[str]=0.0 , lowerCAmelCase__ : List[Any]=0.02 , lowerCAmelCase__ : List[str]=1e-1_2 , lowerCAmelCase__ : Union[str, Any]=224 , lowerCAmelCase__ : Dict=16 , lowerCAmelCase__ : Optional[Any]=3 , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : str=16 , lowerCAmelCase__ : Union[str, Any]=512 , lowerCAmelCase__ : Optional[Any]=8 , lowerCAmelCase__ : Optional[Any]=2048 , lowerCAmelCase__ : Optional[int]=0.75 , lowerCAmelCase__ : List[str]=False , **lowerCAmelCase__ : int , ) -> Dict: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = qkv_bias _UpperCamelCase = decoder_num_attention_heads _UpperCamelCase = decoder_hidden_size _UpperCamelCase = decoder_num_hidden_layers _UpperCamelCase = decoder_intermediate_size _UpperCamelCase = mask_ratio _UpperCamelCase = norm_pix_loss
98
"""simple docstring""" from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( '''The RoBERTa Model transformer with early exiting (DeeRoBERTa). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : str = RobertaEmbeddings(_lowercase ) self.init_weights() @add_start_docstrings( '''RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Optional[Any] = config.num_labels snake_case_ : Dict = config.num_hidden_layers snake_case_ : str = DeeRobertaModel(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : List[str] = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> Tuple: '''simple docstring''' snake_case_ : Any = self.num_layers try: snake_case_ : int = self.roberta( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) snake_case_ : str = outputs[1] snake_case_ : Union[str, Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : Dict = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : List[Any] = e.message snake_case_ : Union[str, Any] = e.exit_layer snake_case_ : Dict = outputs[0] if not self.training: snake_case_ : Dict = entropy(_lowercase ) snake_case_ : Optional[int] = [] snake_case_ : Union[str, Any] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Dict = MSELoss() snake_case_ : Dict = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Union[str, Any] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : int = [] for highway_exit in outputs[-1]: snake_case_ : Tuple = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Optional[Any] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Optional[int] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : Dict = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : List[str] = (loss,) + outputs if not self.training: snake_case_ : Optional[Any] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : Tuple = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
58
0
from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("""repo_id""" , ["""canonical_dataset_name""", """org-name/dataset-name"""] ) @pytest.mark.parametrize("""path""" , ["""filename.csv""", """filename with blanks.csv"""] ) @pytest.mark.parametrize("""revision""" , [None, """v2"""] ) def a (lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): __a = hf_hub_url(repo_id=lowerCAmelCase__ , path=lowerCAmelCase__ , revision=lowerCAmelCase__ ) assert url == f'''https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(lowerCAmelCase__ )}'''
99
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : list[int] ): '''simple docstring''' if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : list[int] , __UpperCamelCase : int ): '''simple docstring''' if curr_ind == len(__UpperCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(__UpperCamelCase ) ): if valid_connection(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): # Insert current vertex into path as next transition snake_case_ : List[str] = next_ver # Validate created path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , curr_ind + 1 ): return True # Backtrack snake_case_ : Tuple = -1 return False def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int = 0 ): '''simple docstring''' snake_case_ : Tuple = [-1] * (len(__UpperCamelCase ) + 1) # initialize start and end of path with starting index snake_case_ : Optional[int] = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , 1 ) else []
58
0
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand _A : str = ( """4S 3H 2C 7S 5H""", """9D 8H 2C 6S 7H""", """2D 6D 9D TH 7D""", """TC 8C 2S JH 6C""", """JH 8S TH AH QH""", """TS KS 5S 9S AC""", """KD 6S 9D TH AD""", """KS 8D 4D 9S 4S""", # pair """8C 4S KH JS 4D""", # pair """QH 8H KD JH 8S""", # pair """KC 4H KS 2H 8D""", # pair """KD 4S KC 3H 8S""", # pair """AH 8S AS KC JH""", # pair """3H 4C 4H 3S 2H""", # 2 pairs """5S 5D 2C KH KH""", # 2 pairs """3C KH 5D 5S KH""", # 2 pairs """AS 3C KH AD KH""", # 2 pairs """7C 7S 3S 7H 5S""", # 3 of a kind """7C 7S KH 2H 7H""", # 3 of a kind """AC KH QH AH AS""", # 3 of a kind """2H 4D 3C AS 5S""", # straight (low ace) """3C 5C 4C 2C 6H""", # straight """6S 8S 7S 5H 9H""", # straight """JS QS 9H TS KH""", # straight """QC KH TS JS AH""", # straight (high ace) """8C 9C 5C 3C TC""", # flush """3S 8S 9S 5S KS""", # flush """4C 5C 9C 8C KC""", # flush """JH 8H AH KH QH""", # flush """3D 2H 3H 2C 2D""", # full house """2H 2C 3S 3H 3D""", # full house """KH KC 3S 3H 3D""", # full house """JC 6H JS JD JH""", # 4 of a kind """JC 7H JS JD JH""", # 4 of a kind """JC KH JS JD JH""", # 4 of a kind """2S AS 4S 5S 3S""", # straight flush (low ace) """2D 6D 3D 4D 5D""", # straight flush """5C 6C 3C 7C 4C""", # straight flush """JH 9H TH KH QH""", # straight flush """JH AH TH KH QH""", # royal flush (high ace straight flush) ) _A : List[Any] = ( ("""2H 3H 4H 5H 6H""", """KS AS TS QS JS""", """Loss"""), ("""2H 3H 4H 5H 6H""", """AS AD AC AH JD""", """Win"""), ("""AS AH 2H AD AC""", """JS JD JC JH 3D""", """Win"""), ("""2S AH 2H AS AC""", """JS JD JC JH AD""", """Loss"""), ("""2S AH 2H AS AC""", """2H 3H 5H 6H 7H""", """Win"""), ("""AS 3S 4S 8S 2S""", """2H 3H 5H 6H 7H""", """Win"""), ("""2H 3H 5H 6H 7H""", """2S 3H 4H 5S 6C""", """Win"""), ("""2S 3H 4H 5S 6C""", """3D 4C 5H 6H 2S""", """Tie"""), ("""2S 3H 4H 5S 6C""", """AH AC 5H 6H AS""", """Win"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H AS""", """Loss"""), ("""2S 2H 4H 5S 4C""", """AH AC 5H 6H 7S""", """Win"""), ("""6S AD 7H 4S AS""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S AH 4H 5S KC""", """AH AC 5H 6H 7S""", """Loss"""), ("""2S 3H 6H 7S 9C""", """7H 3C TH 6H 9S""", """Loss"""), ("""4S 5H 6H TS AC""", """3S 5H 6H TS AC""", """Win"""), ("""2S AH 4H 5S 6C""", """AD 4C 5H 6H 2C""", """Tie"""), ("""AS AH 3H AD AC""", """AS AH 2H AD AC""", """Win"""), ("""AH AC 5H 5C QS""", """AH AC 5H 5C KS""", """Loss"""), ("""AH AC 5H 5C QS""", """KH KC 5H 5C QS""", """Win"""), ("""7C 7S KH 2H 7H""", """3C 3S AH 2H 3H""", """Win"""), ("""3C 3S AH 2H 3H""", """7C 7S KH 2H 7H""", """Loss"""), ("""6H 5H 4H 3H 2H""", """5H 4H 3H 2H AH""", """Win"""), ("""5H 4H 3H 2H AH""", """5H 4H 3H 2H AH""", """Tie"""), ("""5H 4H 3H 2H AH""", """6H 5H 4H 3H 2H""", """Loss"""), ("""AH AD KS KC AC""", """AH KD KH AC KC""", """Win"""), ("""2H 4D 3C AS 5S""", """2H 4D 3C 6S 5S""", """Loss"""), ("""2H 3S 3C 3H 2S""", """3S 3C 2S 2H 2D""", """Win"""), ("""4D 6D 5D 2D JH""", """3S 8S 3H TC KH""", """Loss"""), ("""4S 6C 8S 3S 7S""", """AD KS 2D 7D 7C""", """Loss"""), ("""6S 4C 7H 8C 3H""", """5H JC AH 9D 9C""", """Loss"""), ("""9D 9H JH TC QH""", """3C 2S JS 5C 7H""", """Win"""), ("""2H TC 8S AD 9S""", """4H TS 7H 2C 5C""", """Win"""), ("""9D 3S 2C 7S 7C""", """JC TD 3C TC 9H""", """Loss"""), ) _A : Optional[int] = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", True), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", False), ("""AS 3S 4S 8S 2S""", True), ) _A : List[Any] = ( ("""2H 3H 4H 5H 6H""", True), ("""AS AH 2H AD AC""", False), ("""2H 3H 5H 6H 7H""", False), ("""KS AS TS QS JS""", True), ("""8H 9H QS JS TH""", True), ) _A : List[str] = ( ("""2H 4D 3C AS 5S""", True, [5, 4, 3, 2, 14]), ("""2H 5D 3C AS 5S""", False, [14, 5, 5, 3, 2]), ("""JH QD KC AS TS""", False, [14, 13, 12, 11, 10]), ("""9D 3S 2C 7S 7C""", False, [9, 7, 7, 3, 2]), ) _A : Union[str, Any] = ( ("""JH AH TH KH QH""", 0), ("""JH 9H TH KH QH""", 0), ("""JC KH JS JD JH""", 7), ("""KH KC 3S 3H 3D""", 6), ("""8C 9C 5C 3C TC""", 0), ("""JS QS 9H TS KH""", 0), ("""7C 7S KH 2H 7H""", 3), ("""3C KH 5D 5S KH""", 2), ("""QH 8H KD JH 8S""", 1), ("""2D 6D 9D TH 7D""", 0), ) _A : List[str] = ( ("""JH AH TH KH QH""", 23), ("""JH 9H TH KH QH""", 22), ("""JC KH JS JD JH""", 21), ("""KH KC 3S 3H 3D""", 20), ("""8C 9C 5C 3C TC""", 19), ("""JS QS 9H TS KH""", 18), ("""7C 7S KH 2H 7H""", 17), ("""3C KH 5D 5S KH""", 16), ("""QH 8H KD JH 8S""", 15), ("""2D 6D 9D TH 7D""", 14), ) def __snake_case ( ) -> str: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = randrange(len(lowerCAmelCase_ ) ), randrange(len(lowerCAmelCase_ ) ) SCREAMING_SNAKE_CASE__ = ['''Loss''', '''Tie''', '''Win'''][(play >= oppo) + (play > oppo)] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def __snake_case ( lowerCAmelCase_ = 1_0_0 ) -> List[Any]: return (generate_random_hand() for _ in range(lowerCAmelCase_ )) @pytest.mark.parametrize('''hand, expected''' , lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Union[str, Any]: assert PokerHand(lowerCAmelCase_ )._is_flush() == expected @pytest.mark.parametrize('''hand, expected''' , lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: assert PokerHand(lowerCAmelCase_ )._is_straight() == expected @pytest.mark.parametrize('''hand, expected, card_values''' , lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = PokerHand(lowerCAmelCase_ ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('''hand, expected''' , lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Dict: assert PokerHand(lowerCAmelCase_ )._is_same_kind() == expected @pytest.mark.parametrize('''hand, expected''' , lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Union[str, Any]: assert PokerHand(lowerCAmelCase_ )._hand_type == expected @pytest.mark.parametrize('''hand, other, expected''' , lowerCAmelCase_ ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Tuple: assert PokerHand(lowerCAmelCase_ ).compare_with(PokerHand(lowerCAmelCase_ ) ) == expected @pytest.mark.parametrize('''hand, other, expected''' , generate_random_hands() ) def __snake_case ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[Any]: assert PokerHand(lowerCAmelCase_ ).compare_with(PokerHand(lowerCAmelCase_ ) ) == expected def __snake_case ( ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = [PokerHand(lowerCAmelCase_ ) for hand in SORTED_HANDS] SCREAMING_SNAKE_CASE__ = poker_hands.copy() shuffle(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = chain(sorted(lowerCAmelCase_ ) ) for index, hand in enumerate(lowerCAmelCase_ ): assert hand == poker_hands[index] def __snake_case ( ) -> Tuple: # Test that five high straights are compared correctly. SCREAMING_SNAKE_CASE__ = [PokerHand('''2D AC 3H 4H 5S''' ), PokerHand('''2S 3H 4H 5S 6C''' )] pokerhands.sort(reverse=lowerCAmelCase_ ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def __snake_case ( ) -> Tuple: # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. SCREAMING_SNAKE_CASE__ = PokerHand('''2C 4S AS 3D 5C''' ) SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = [5, 4, 3, 2, 1_4] for _ in range(1_0 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def __snake_case ( ) -> str: # Problem number 54 from Project Euler # Testing from poker_hands.txt file SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = os.path.abspath(os.path.dirname(lowerCAmelCase_ ) ) SCREAMING_SNAKE_CASE__ = os.path.join(lowerCAmelCase_ , '''poker_hands.txt''' ) with open(lowerCAmelCase_ ) as file_hand: for line in file_hand: SCREAMING_SNAKE_CASE__ = line[:1_4].strip() SCREAMING_SNAKE_CASE__ = line[1_5:].strip() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = PokerHand(lowerCAmelCase_ ), PokerHand(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__ = player.compare_with(lowerCAmelCase_ ) if output == "Win": answer += 1 assert answer == 3_7_6
100
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = ['''image_processor''', '''tokenizer'''] _lowerCamelCase = '''BlipImageProcessor''' _lowerCamelCase = '''AutoTokenizer''' def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__(_lowercase , _lowercase ) # add QFormer tokenizer snake_case_ : List[str] = qformer_tokenizer def __call__( self , _lowercase = None , _lowercase = None , _lowercase = True , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = 0 , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = True , _lowercase = None , **_lowercase , ) -> BatchFeature: '''simple docstring''' if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) snake_case_ : Optional[Any] = BatchFeature() if text is not None: snake_case_ : List[str] = self.tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) encoding.update(_lowercase ) snake_case_ : Union[str, Any] = self.qformer_tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) snake_case_ : List[str] = qformer_text_encoding.pop("""input_ids""" ) snake_case_ : Union[str, Any] = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: snake_case_ : Tuple = self.image_processor(_lowercase , return_tensors=_lowercase ) encoding.update(_lowercase ) return encoding def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> List[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> Dict: '''simple docstring''' return self.tokenizer.decode(*_lowercase , **_lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' snake_case_ : str = self.tokenizer.model_input_names snake_case_ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def UpperCAmelCase__ ( self , _lowercase , **_lowercase ) -> Optional[int]: '''simple docstring''' if os.path.isfile(_lowercase ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(_lowercase , exist_ok=_lowercase ) snake_case_ : int = os.path.join(_lowercase , """qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(_lowercase ) return super().save_pretrained(_lowercase , **_lowercase ) @classmethod def UpperCAmelCase__ ( cls , _lowercase , **_lowercase ) -> int: '''simple docstring''' snake_case_ : List[str] = AutoTokenizer.from_pretrained(_lowercase , subfolder="""qformer_tokenizer""" ) snake_case_ : Union[str, Any] = cls._get_arguments_from_pretrained(_lowercase , **_lowercase ) args.append(_lowercase ) return cls(*_lowercase )
58
0
from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) lowerCAmelCase__ : Optional[int] =_symbol_database.Default() lowerCAmelCase__ : Optional[int] =_descriptor_pool.Default().AddSerializedFile( b'\n\x19sentencepiece_model.proto\x12\rsentencepiece"\x80\x0c\n\x0bTrainerSpec\x12\r\n\x05input\x18\x01 \x03(\t\x12\x14\n\x0cinput_format\x18\x07 \x01(\t\x12\x14\n\x0cmodel_prefix\x18\x02 \x01(\t\x12\x41\n\nmodel_type\x18\x03 \x01(\x0e\x32$.sentencepiece.TrainerSpec.ModelType:\x07UNIGRAM\x12\x18\n\nvocab_size\x18\x04 \x01(\x05:\x04\x38\x30\x30\x30\x12\x17\n\x0f\x61\x63\x63\x65pt_language\x18\x05 \x03(\t\x12 \n\x15self_test_sample_size\x18\x06 \x01(\x05:\x01\x30\x12*\n\x1b\x65nable_differential_privacy\x18\x32 \x01(\x08:\x05\x66\x61lse\x12+\n differential_privacy_noise_level\x18\x33 \x01(\x02:\x01\x30\x12\x32\n\'differential_privacy_clipping_threshold\x18\x34 \x01(\x04:\x01\x30\x12"\n\x12\x63haracter_coverage\x18\n \x01(\x02:\x06\x30.9995\x12\x1e\n\x13input_sentence_size\x18\x0b \x01(\x04:\x01\x30\x12$\n\x16shuffle_input_sentence\x18\x13 \x01(\x08:\x04true\x12 \n\x14mining_sentence_size\x18\x0c \x01(\x05\x42\x02\x18\x01\x12"\n\x16training_sentence_size\x18\r \x01(\x05\x42\x02\x18\x01\x12(\n\x17seed_sentencepiece_size\x18\x0e \x01(\x05:\x07\x31\x30\x30\x30\x30\x30\x30\x12\x1e\n\x10shrinking_factor\x18\x0f \x01(\x02:\x04\x30.75\x12!\n\x13max_sentence_length\x18\x12 \x01(\x05:\x04\x34\x31\x39\x32\x12\x17\n\x0bnum_threads\x18\x10 \x01(\x05:\x02\x31\x36\x12\x1d\n\x12num_sub_iterations\x18\x11 \x01(\x05:\x01\x32\x12$\n\x18max_sentencepiece_length\x18\x14 \x01(\x05:\x02\x31\x36\x12%\n\x17split_by_unicode_script\x18\x15 \x01(\x08:\x04true\x12\x1d\n\x0fsplit_by_number\x18\x17 \x01(\x08:\x04true\x12!\n\x13split_by_whitespace\x18\x16 \x01(\x08:\x04true\x12)\n\x1atreat_whitespace_as_suffix\x18\x18 \x01(\x08:\x05\x66\x61lse\x12+\n\x1c\x61llow_whitespace_only_pieces\x18\x1a \x01(\x08:\x05\x66\x61lse\x12\x1b\n\x0csplit_digits\x18\x19 \x01(\x08:\x05\x66\x61lse\x12#\n\x19pretokenization_delimiter\x18\x35 \x01(\t:\x00\x12\x17\n\x0f\x63ontrol_symbols\x18\x1e \x03(\t\x12\x1c\n\x14user_defined_symbols\x18\x1f \x03(\t\x12\x16\n\x0erequired_chars\x18$ \x01(\t\x12\x1c\n\rbyte_fallback\x18# \x01(\x08:\x05\x66\x61lse\x12+\n\x1dvocabulary_output_piece_score\x18 \x01(\x08:\x04true\x12\x1e\n\x10hard_vocab_limit\x18! \x01(\x08:\x04true\x12\x1c\n\ruse_all_vocab\x18" \x01(\x08:\x05\x66\x61lse\x12\x11\n\x06unk_id\x18( \x01(\x05:\x01\x30\x12\x11\n\x06\x62os_id\x18) \x01(\x05:\x01\x31\x12\x11\n\x06\x65os_id\x18* \x01(\x05:\x01\x32\x12\x12\n\x06pad_id\x18+ \x01(\x05:\x02-1\x12\x18\n\tunk_piece\x18- \x01(\t:\x05<unk>\x12\x16\n\tbos_piece\x18. \x01(\t:\x03<s>\x12\x17\n\teos_piece\x18/ \x01(\t:\x04</s>\x12\x18\n\tpad_piece\x18\x30 \x01(\t:\x05<pad>\x12\x1a\n\x0bunk_surface\x18, \x01(\t:\x05 \xe2\x81\x87 \x12+\n\x1ctrain_extremely_large_corpus\x18\x31 \x01(\x08:\x05\x66\x61lse"5\n\tModelType\x12\x0b\n\x07UNIGRAM\x10\x01\x12\x07\n\x03\x42PE\x10\x02\x12\x08\n\x04WORD\x10\x03\x12\x08\n\x04\x43HAR\x10\x04*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"\xd1\x01\n\x0eNormalizerSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x1c\n\x14precompiled_charsmap\x18\x02 \x01(\x0c\x12\x1e\n\x10\x61\x64\x64_dummy_prefix\x18\x03 \x01(\x08:\x04true\x12&\n\x18remove_extra_whitespaces\x18\x04 \x01(\x08:\x04true\x12 \n\x12\x65scape_whitespaces\x18\x05 \x01(\x08:\x04true\x12\x1e\n\x16normalization_rule_tsv\x18\x06 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"y\n\x0cSelfTestData\x12\x33\n\x07samples\x18\x01 \x03(\x0b\x32".sentencepiece.SelfTestData.Sample\x1a)\n\x06Sample\x12\r\n\x05input\x18\x01 \x01(\t\x12\x10\n\x08\x65xpected\x18\x02 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"\xfe\x03\n\nModelProto\x12\x37\n\x06pieces\x18\x01 \x03(\x0b\x32\'.sentencepiece.ModelProto.SentencePiece\x12\x30\n\x0ctrainer_spec\x18\x02 \x01(\x0b\x32\x1a.sentencepiece.TrainerSpec\x12\x36\n\x0fnormalizer_spec\x18\x03 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x12\x33\n\x0eself_test_data\x18\x04 \x01(\x0b\x32\x1b.sentencepiece.SelfTestData\x12\x38\n\x11\x64\x65normalizer_spec\x18\x05 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x1a\xd2\x01\n\rSentencePiece\x12\r\n\x05piece\x18\x01 \x01(\t\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x42\n\x04type\x18\x03 \x01(\x0e\x32,.sentencepiece.ModelProto.SentencePiece.Type:\x06NORMAL"T\n\x04Type\x12\n\n\x06NORMAL\x10\x01\x12\x0b\n\x07UNKNOWN\x10\x02\x12\x0b\n\x07\x43ONTROL\x10\x03\x12\x10\n\x0cUSER_DEFINED\x10\x04\x12\x08\n\x04\x42YTE\x10\x06\x12\n\n\x06UNUSED\x10\x05*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\x42\x02H\x03' ) lowerCAmelCase__ : Any =globals() _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) _builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'sentencepiece_model_pb2', _globals) if _descriptor._USE_C_DESCRIPTORS is False: lowerCAmelCase__ : List[Any] =None lowerCAmelCase__ : Any =b'H\003' # (generated by protobuf compiler, but `_TRAINERSPEC` is not defined) # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._serialized_options = b"\030\001" # _TRAINERSPEC.fields_by_name["training_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["training_sentence_size"]._serialized_options = b"\030\001" lowerCAmelCase__ : str =45 lowerCAmelCase__ : List[str] =15_81 lowerCAmelCase__ : Optional[int] =15_17 lowerCAmelCase__ : Any =15_70 lowerCAmelCase__ : Optional[Any] =15_84 lowerCAmelCase__ : List[str] =17_93 lowerCAmelCase__ : int =17_95 lowerCAmelCase__ : Optional[int] =19_16 lowerCAmelCase__ : str =18_64 lowerCAmelCase__ : int =19_05 lowerCAmelCase__ : Union[str, Any] =19_19 lowerCAmelCase__ : Any =24_29 lowerCAmelCase__ : Union[str, Any] =22_08 lowerCAmelCase__ : List[Any] =24_18 lowerCAmelCase__ : Optional[int] =23_23 lowerCAmelCase__ : int =24_07 # @@protoc_insertion_point(module_scope)
101
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCAmelCase : List[Any] = { '''configuration_vivit''': ['''VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VivitConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = ['''VivitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = [ '''VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VivitModel''', '''VivitPreTrainedModel''', '''VivitForVideoClassification''', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
58
0
"""simple docstring""" from itertools import permutations def UpperCamelCase (SCREAMING_SNAKE_CASE ): if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False UpperCamelCase : Optional[int] = [7, 11, 13, 17] for i, test in enumerate(SCREAMING_SNAKE_CASE ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def UpperCamelCase (SCREAMING_SNAKE_CASE = 10 ): return sum( int("""""".join(map(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) ) for num in permutations(range(SCREAMING_SNAKE_CASE ) ) if is_substring_divisible(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": print(f'''{solution() = }''')
102
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() __lowerCAmelCase : Dict = logging.get_logger(__name__) def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Dict , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : List[str] = WavaVecaForSequenceClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : int = downstream_dict["""projector.weight"""] snake_case_ : Optional[int] = downstream_dict["""projector.bias"""] snake_case_ : List[Any] = downstream_dict["""model.post_net.linear.weight"""] snake_case_ : Union[str, Any] = downstream_dict["""model.post_net.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Tuple , __UpperCamelCase : str ): '''simple docstring''' snake_case_ : int = WavaVecaForAudioFrameClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""model.linear.weight"""] snake_case_ : int = downstream_dict["""model.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[str] , __UpperCamelCase : List[Any] ): '''simple docstring''' snake_case_ : Optional[int] = WavaVecaForXVector.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""connector.weight"""] snake_case_ : str = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): snake_case_ : Dict = downstream_dict[ F'model.framelevel_feature_extractor.module.{i}.kernel.weight' ] snake_case_ : int = downstream_dict[F'model.framelevel_feature_extractor.module.{i}.kernel.bias'] snake_case_ : str = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] snake_case_ : Any = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] snake_case_ : List[str] = downstream_dict["""objective.W"""] return model @torch.no_grad() def __lowerCAmelCase ( __UpperCamelCase : List[Any] , __UpperCamelCase : Any , __UpperCamelCase : List[str] , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : Any = torch.load(__UpperCamelCase , map_location="""cpu""" ) snake_case_ : Any = checkpoint["""Downstream"""] snake_case_ : Optional[Any] = WavaVecaConfig.from_pretrained(__UpperCamelCase ) snake_case_ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained( __UpperCamelCase , return_attention_mask=__UpperCamelCase , do_normalize=__UpperCamelCase ) snake_case_ : Optional[Any] = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): snake_case_ : Tuple = convert_classification(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForAudioFrameClassification""" ): snake_case_ : Union[str, Any] = convert_diarization(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForXVector""" ): snake_case_ : List[str] = convert_xvector(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: raise NotImplementedError(F'S3PRL weights conversion is not supported for {arch}' ) if hf_config.use_weighted_layer_sum: snake_case_ : List[Any] = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(__UpperCamelCase ) hf_model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __lowerCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--base_model_name''', default=None, type=str, help='''Name of the huggingface pretrained base model.''' ) parser.add_argument('''--config_path''', default=None, type=str, help='''Path to the huggingface classifier config.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to the s3prl checkpoint.''') parser.add_argument('''--model_dump_path''', default=None, type=str, help='''Path to the final converted model.''') __lowerCAmelCase : Dict = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
58
0
"""simple docstring""" from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): def __init__( self : Optional[int] , __lowerCamelCase : Callable , __lowerCamelCase : Optional[Features] = None , __lowerCamelCase : str = None , __lowerCamelCase : bool = False , __lowerCamelCase : bool = False , __lowerCamelCase : Optional[dict] = None , __lowerCamelCase : Optional[int] = None , **__lowerCamelCase : Union[str, Any] , ): """simple docstring""" super().__init__( features=__lowerCamelCase , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase , streaming=__lowerCamelCase , num_proc=__lowerCamelCase , **__lowerCamelCase , ) _snake_case = Generator( cache_dir=__lowerCamelCase , features=__lowerCamelCase , generator=__lowerCamelCase , gen_kwargs=__lowerCamelCase , **__lowerCamelCase , ) def __UpperCAmelCase ( self : Optional[int] ): """simple docstring""" # Build iterable dataset if self.streaming: _snake_case = self.builder.as_streaming_dataset(split='''train''' ) # Build regular (map-style) dataset else: _snake_case = None _snake_case = None _snake_case = None _snake_case = None self.builder.download_and_prepare( download_config=__lowerCamelCase , download_mode=__lowerCamelCase , verification_mode=__lowerCamelCase , base_path=__lowerCamelCase , num_proc=self.num_proc , ) _snake_case = self.builder.as_dataset( split='''train''' , verification_mode=__lowerCamelCase , in_memory=self.keep_in_memory ) return dataset
103
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __lowerCAmelCase : Tuple = logging.get_logger(__name__) __lowerCAmelCase : int = {'''vocab_file''': '''vocab.txt'''} __lowerCAmelCase : Union[str, Any] = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } __lowerCAmelCase : Optional[Any] = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } __lowerCAmelCase : Any = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = PRETRAINED_INIT_CONFIGURATION _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase = ConvBertTokenizer def __init__( self , _lowercase=None , _lowercase=None , _lowercase=True , _lowercase="[UNK]" , _lowercase="[SEP]" , _lowercase="[PAD]" , _lowercase="[CLS]" , _lowercase="[MASK]" , _lowercase=True , _lowercase=None , **_lowercase , ) -> Union[str, Any]: '''simple docstring''' super().__init__( _lowercase , tokenizer_file=_lowercase , do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , tokenize_chinese_chars=_lowercase , strip_accents=_lowercase , **_lowercase , ) snake_case_ : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowercase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowercase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowercase ) != tokenize_chinese_chars ): snake_case_ : Optional[int] = getattr(_lowercase , normalizer_state.pop("""type""" ) ) snake_case_ : Dict = do_lower_case snake_case_ : str = strip_accents snake_case_ : Optional[Any] = tokenize_chinese_chars snake_case_ : int = normalizer_class(**_lowercase ) snake_case_ : Optional[int] = do_lower_case def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> int: '''simple docstring''' snake_case_ : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> List[int]: '''simple docstring''' snake_case_ : int = [self.sep_token_id] snake_case_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[int] = self._tokenizer.model.save(_lowercase , name=_lowercase ) return tuple(_lowercase )
58
0
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { """salesforce/blip2-opt-2.7b""": """https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json""", } class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" A__ : Optional[int] = "blip_2_vision_model" def __init__( self , SCREAMING_SNAKE_CASE__=1408 , SCREAMING_SNAKE_CASE__=6144 , SCREAMING_SNAKE_CASE__=39 , SCREAMING_SNAKE_CASE__=16 , SCREAMING_SNAKE_CASE__=224 , SCREAMING_SNAKE_CASE__=14 , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=0.0_0_0_0_1 , SCREAMING_SNAKE_CASE__=0.0 , SCREAMING_SNAKE_CASE__=1e-10 , SCREAMING_SNAKE_CASE__=True , **SCREAMING_SNAKE_CASE__ , ) -> Optional[Any]: super().__init__(**SCREAMING_SNAKE_CASE__ ) A__ = hidden_size A__ = intermediate_size A__ = num_hidden_layers A__ = num_attention_heads A__ = patch_size A__ = image_size A__ = initializer_range A__ = attention_dropout A__ = layer_norm_eps A__ = hidden_act A__ = qkv_bias @classmethod def snake_case__ ( cls , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(SCREAMING_SNAKE_CASE__ ) A__ , A__ = cls.get_config_dict(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": A__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" A__ : List[Any] = "blip_2_qformer" def __init__( self , SCREAMING_SNAKE_CASE__=30522 , SCREAMING_SNAKE_CASE__=768 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=12 , SCREAMING_SNAKE_CASE__=3072 , SCREAMING_SNAKE_CASE__="gelu" , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=0.1 , SCREAMING_SNAKE_CASE__=512 , SCREAMING_SNAKE_CASE__=0.0_2 , SCREAMING_SNAKE_CASE__=1e-12 , SCREAMING_SNAKE_CASE__=0 , SCREAMING_SNAKE_CASE__="absolute" , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=1408 , **SCREAMING_SNAKE_CASE__ , ) -> List[Any]: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = hidden_act A__ = intermediate_size A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = initializer_range A__ = layer_norm_eps A__ = position_embedding_type A__ = cross_attention_frequency A__ = encoder_hidden_size @classmethod def snake_case__ ( cls , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> "PretrainedConfig": cls._set_token_in_kwargs(SCREAMING_SNAKE_CASE__ ) A__ , A__ = cls.get_config_dict(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": A__ = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) class UpperCamelCase__ ( _lowerCAmelCase ): """simple docstring""" A__ : Tuple = "blip-2" A__ : Optional[Any] = True def __init__( self , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=None , SCREAMING_SNAKE_CASE__=32 , **SCREAMING_SNAKE_CASE__ ) -> List[Any]: super().__init__(**SCREAMING_SNAKE_CASE__ ) if vision_config is None: A__ = {} logger.info("vision_config is None. initializing the Blip2VisionConfig with default values." ) if qformer_config is None: A__ = {} logger.info("qformer_config is None. Initializing the Blip2QFormerConfig with default values." ) if text_config is None: A__ = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) A__ = BlipaVisionConfig(**SCREAMING_SNAKE_CASE__ ) A__ = BlipaQFormerConfig(**SCREAMING_SNAKE_CASE__ ) A__ = text_config["model_type"] if "model_type" in text_config else "opt" A__ = CONFIG_MAPPING[text_model_type](**SCREAMING_SNAKE_CASE__ ) A__ = self.text_config.tie_word_embeddings A__ = self.text_config.is_encoder_decoder A__ = num_query_tokens A__ = self.vision_config.hidden_size A__ = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES A__ = 1.0 A__ = 0.0_2 @classmethod def snake_case__ ( cls , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) -> Union[str, Any]: return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **SCREAMING_SNAKE_CASE__ , ) def snake_case__ ( self ) -> List[Any]: A__ = copy.deepcopy(self.__dict__ ) A__ = self.vision_config.to_dict() A__ = self.qformer_config.to_dict() A__ = self.text_config.to_dict() A__ = self.__class__.model_type return output
104
"""simple docstring""" import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" @register_to_config def __init__( self , _lowercase = 1_2_8 , _lowercase = 2_5_6 , _lowercase = 2000.0 , _lowercase = 7_6_8 , _lowercase = 1_2 , _lowercase = 1_2 , _lowercase = 6_4 , _lowercase = 2_0_4_8 , _lowercase = 0.1 , ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Optional[Any] = nn.Sequential( nn.Linear(_lowercase , d_model * 4 , bias=_lowercase ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_lowercase ) , nn.SiLU() , ) snake_case_ : Any = nn.Embedding(_lowercase , _lowercase ) snake_case_ : Union[str, Any] = False snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Union[str, Any] = nn.Dropout(p=_lowercase ) snake_case_ : Tuple = nn.ModuleList() for lyr_num in range(_lowercase ): # FiLM conditional T5 decoder snake_case_ : Union[str, Any] = DecoderLayer(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) self.decoders.append(_lowercase ) snake_case_ : List[Any] = TaLayerNorm(_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(p=_lowercase ) snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def UpperCAmelCase__ ( self , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ , snake_case_ , snake_case_ : str = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. snake_case_ : Optional[int] = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) snake_case_ : int = self.conditioning_emb(_lowercase ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) snake_case_ : Tuple = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. snake_case_ : Dict = torch.broadcast_to( torch.arange(_lowercase , device=decoder_input_tokens.device ) , (batch, seq_length) , ) snake_case_ : Tuple = self.position_encoding(_lowercase ) snake_case_ : Optional[Any] = self.continuous_inputs_projection(_lowercase ) inputs += position_encodings snake_case_ : List[Any] = self.dropout(_lowercase ) # decoder: No padding present. snake_case_ : Tuple = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. snake_case_ : int = [(x, self.encoder_decoder_mask(_lowercase , _lowercase )) for x, y in encodings_and_masks] # cross attend style: concat encodings snake_case_ : Optional[Any] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) snake_case_ : str = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: snake_case_ : int = lyr( _lowercase , conditioning_emb=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , )[0] snake_case_ : int = self.decoder_norm(_lowercase ) snake_case_ : Union[str, Any] = self.post_dropout(_lowercase ) snake_case_ : int = self.spec_out(_lowercase ) return spec_out class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase=1E-6 ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Any = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase ) ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = self.layer[0]( _lowercase , conditioning_emb=_lowercase , attention_mask=_lowercase , ) if encoder_hidden_states is not None: snake_case_ : Tuple = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) snake_case_ : str = self.layer[1]( _lowercase , key_value_states=_lowercase , attention_mask=_lowercase , ) # Apply Film Conditional Feed Forward layer snake_case_ : Any = self.layer[-1](_lowercase , _lowercase ) return (hidden_states,) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> str: '''simple docstring''' super().__init__() snake_case_ : Any = TaLayerNorm(_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Union[str, Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : List[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : str = self.FiLMLayer(_lowercase , _lowercase ) # Self-attention block snake_case_ : List[Any] = self.attention(_lowercase ) snake_case_ : List[str] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__() snake_case_ : List[Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : Union[str, Any] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) snake_case_ : Optional[Any] = self.attention( _lowercase , encoder_hidden_states=_lowercase , attention_mask=attention_mask.squeeze(1 ) , ) snake_case_ : Any = hidden_states + self.dropout(_lowercase ) return layer_output class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Tuple = TaDenseGatedActDense(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Optional[int] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Tuple = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> str: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : Optional[int] = self.film(_lowercase , _lowercase ) snake_case_ : int = self.DenseReluDense(_lowercase ) snake_case_ : Optional[Any] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__() snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Any = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : int = nn.Dropout(_lowercase ) snake_case_ : Optional[int] = NewGELUActivation() def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' snake_case_ : str = self.act(self.wi_a(_lowercase ) ) snake_case_ : Dict = self.wi_a(_lowercase ) snake_case_ : Any = hidden_gelu * hidden_linear snake_case_ : List[Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.wo(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1E-6 ) -> str: '''simple docstring''' super().__init__() snake_case_ : Union[str, Any] = nn.Parameter(torch.ones(_lowercase ) ) snake_case_ : int = eps def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=_lowercase ) snake_case_ : Any = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: snake_case_ : str = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def UpperCAmelCase__ ( self , _lowercase ) -> torch.Tensor: '''simple docstring''' return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.04_4715 * torch.pow(_lowercase , 3.0 )) )) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Any: '''simple docstring''' super().__init__() snake_case_ : List[Any] = nn.Linear(_lowercase , out_features * 2 , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = self.scale_bias(_lowercase ) snake_case_ , snake_case_ : Any = torch.chunk(_lowercase , 2 , -1 ) snake_case_ : Optional[Any] = x * (1 + scale) + shift return x
58
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=lowerCamelCase_ ) class lowerCAmelCase_ ( lowerCamelCase_ ): __a : str = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) __a : ClassVar[Features] = Features({"image": Image()} ) __a : ClassVar[Features] = Features({"labels": ClassLabel} ) __a : str = "image" __a : str = "labels" def snake_case ( self ,snake_case__ ): if self.label_column not in features: raise ValueError(F'Column {self.label_column} is not present in features.' ) if not isinstance(features[self.label_column] ,snake_case__ ): raise ValueError(F'Column {self.label_column} is not a ClassLabel.' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = copy.deepcopy(self ) SCREAMING_SNAKE_CASE_ : Tuple = self.label_schema.copy() SCREAMING_SNAKE_CASE_ : int = features[self.label_column] SCREAMING_SNAKE_CASE_ : Tuple = label_schema return task_template @property def snake_case ( self ): return { self.image_column: "image", self.label_column: "labels", }
105
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''roformer''' def __init__( self , _lowercase=5_0_0_0_0 , _lowercase=None , _lowercase=7_6_8 , _lowercase=1_2 , _lowercase=1_2 , _lowercase=3_0_7_2 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=1_5_3_6 , _lowercase=2 , _lowercase=0.02 , _lowercase=1E-12 , _lowercase=0 , _lowercase=False , _lowercase=True , **_lowercase , ) -> int: '''simple docstring''' super().__init__(pad_token_id=_lowercase , **_lowercase ) snake_case_ : str = vocab_size snake_case_ : Any = hidden_size if embedding_size is None else embedding_size snake_case_ : List[str] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : Tuple = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : List[str] = rotary_value snake_case_ : str = use_cache class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ : Any = {0: """batch""", 1: """sequence"""} snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
58
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __snake_case :Dict =logging.get_logger(__name__) __snake_case :Any ={'vocab_file': 'sentencepiece.bpe.model'} __snake_case :Union[str, Any] ={ 'vocab_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model' ), }, } __snake_case :Optional[int] ={ 'moussaKam/mbarthez': 1024, 'moussaKam/barthez': 1024, 'moussaKam/barthez-orangesum-title': 1024, } __snake_case :str ='▁' class lowerCAmelCase__ ( _lowerCamelCase ): A_ : Union[str, Any] = VOCAB_FILES_NAMES A_ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP A_ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ : str = ['input_ids', 'attention_mask'] def __init__( self : str , __UpperCamelCase : Tuple , __UpperCamelCase : List[str]="<s>" , __UpperCamelCase : Optional[int]="</s>" , __UpperCamelCase : Any="</s>" , __UpperCamelCase : int="<s>" , __UpperCamelCase : Optional[Any]="<unk>" , __UpperCamelCase : Dict="<pad>" , __UpperCamelCase : str="<mask>" , __UpperCamelCase : Optional[Dict[str, Any]] = None , **__UpperCamelCase : Optional[Any] , ) -> None: # Mask token behave like a normal word, i.e. include the space before it A = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token A = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , sep_token=__UpperCamelCase , cls_token=__UpperCamelCase , pad_token=__UpperCamelCase , mask_token=__UpperCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCamelCase , ) A = vocab_file A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__UpperCamelCase ) ) A = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} A = len(self.sp_model ) - 1 A = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __UpperCamelCase ( self : Dict , __UpperCamelCase : List[int] , __UpperCamelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] A = [self.cls_token_id] A = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __UpperCamelCase ( self : Any , __UpperCamelCase : List[int] , __UpperCamelCase : Optional[List[int]] = None , __UpperCamelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCamelCase , token_ids_a=__UpperCamelCase , already_has_special_tokens=__UpperCamelCase ) if token_ids_a is None: return [1] + ([0] * len(__UpperCamelCase )) + [1] return [1] + ([0] * len(__UpperCamelCase )) + [1, 1] + ([0] * len(__UpperCamelCase )) + [1] def __UpperCamelCase ( self : List[str] , __UpperCamelCase : List[int] , __UpperCamelCase : Optional[List[int]] = None ) -> List[int]: A = [self.sep_token_id] A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __UpperCamelCase ( self : Optional[Any] ) -> Dict: return len(self.sp_model ) def __UpperCamelCase ( self : Tuple ) -> str: A = {self.convert_ids_to_tokens(__UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __UpperCamelCase ( self : Union[str, Any] , __UpperCamelCase : str ) -> List[str]: return self.sp_model.encode(__UpperCamelCase , out_type=__UpperCamelCase ) def __UpperCamelCase ( self : Dict , __UpperCamelCase : Any ) -> Dict: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] A = self.sp_model.PieceToId(__UpperCamelCase ) return spm_id if spm_id else self.unk_token_id def __UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : int ) -> str: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__UpperCamelCase ) def __UpperCamelCase ( self : Any , __UpperCamelCase : Optional[int] ) -> Optional[int]: A = [] A = '' A = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCamelCase ) + token A = True A = [] else: current_sub_tokens.append(__UpperCamelCase ) A = False out_string += self.sp_model.decode(__UpperCamelCase ) return out_string.strip() def __getstate__( self : Any ) -> int: A = self.__dict__.copy() A = None return state def __setstate__( self : Any , __UpperCamelCase : Dict ) -> int: A = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): A = {} A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __UpperCamelCase ( self : Tuple , __UpperCamelCase : str , __UpperCamelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__UpperCamelCase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return A = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCamelCase , 'wb' ) as fi: A = self.sp_model.serialized_model_proto() fi.write(__UpperCamelCase ) return (out_vocab_file,)
106
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __lowerCAmelCase ( __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Dict = checkpoints.load_tax_checkpoint(__UpperCamelCase ) snake_case_ : Tuple = flatten_dict(__UpperCamelCase ) return flax_params def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] ): '''simple docstring''' snake_case_ : Optional[Any] = {} snake_case_ : List[Any] = { """token_embedder""": """embeddings""", """encoder_norm""": """layernorm""", """kernel""": """weight""", """.out""": """.output""", """scale""": """weight""", """embedders_0.pos_embedding""": """row_embedder.weight""", """embedders_1.pos_embedding""": """column_embedder.weight""", } snake_case_ : Optional[Any] = { """query""": """attention.query""", """key""": """attention.key""", """value""": """attention.value""", """output.dense""": """output""", """encoder_decoder_attention.o""": """encoder_decoder_attention.attention.o""", """pre_self_attention_layer_norm""": """self_attention.layer_norm""", """pre_cross_attention_layer_norm""": """encoder_decoder_attention.layer_norm""", """mlp.""": """mlp.DenseReluDense.""", """pre_mlp_layer_norm""": """mlp.layer_norm""", """self_attention.o""": """self_attention.attention.o""", """decoder.embeddings.embedding""": """decoder.embed_tokens.weight""", """decoder.relpos_bias.rel_embedding""": """decoder.layer.0.self_attention.attention.relative_attention_bias.weight""", """decoder.decoder_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.logits_dense.weight""": """decoder.lm_head.weight""", } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key snake_case_ : List[Any] = """.""".join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): snake_case_ : List[str] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): snake_case_ : Optional[int] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number snake_case_ : Optional[Any] = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Union[str, Any] = new_key.replace("""encoder""" , """encoder.encoder""" ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number snake_case_ : int = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Dict = flax_dict[key] snake_case_ : Tuple = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): snake_case_ : Optional[int] = torch.from_numpy(converted_dict[key].T ) else: snake_case_ : List[Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : List[str]=False ): '''simple docstring''' snake_case_ : Optional[int] = get_flax_param(__UpperCamelCase ) if not use_large: snake_case_ : Optional[int] = PixaStructVisionConfig() snake_case_ : Optional[Any] = PixaStructTextConfig() else: snake_case_ : Tuple = PixaStructVisionConfig( hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_attention_heads=2_4 , num_hidden_layers=1_8 ) snake_case_ : List[str] = PixaStructTextConfig(hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_heads=2_4 , num_layers=1_8 ) snake_case_ : str = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=__UpperCamelCase ) snake_case_ : Optional[int] = PixaStructForConditionalGeneration(__UpperCamelCase ) snake_case_ : str = rename_and_convert_flax_params(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) snake_case_ : Optional[int] = AutoTokenizer.from_pretrained("""ybelkada/test-pix2struct-tokenizer""" ) snake_case_ : int = PixaStructImageProcessor() snake_case_ : str = PixaStructProcessor(image_processor=__UpperCamelCase , tokenizer=__UpperCamelCase ) if use_large: snake_case_ : Optional[Any] = 4_0_9_6 snake_case_ : int = True # mkdir if needed os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) processor.save_pretrained(__UpperCamelCase ) print("""Model saved in {}""".format(__UpperCamelCase ) ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') __lowerCAmelCase : List[Any] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
58
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} _UpperCAmelCase : List[str] = { '''vocab_file''': { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt''' ), } } _UpperCAmelCase : Optional[int] = { '''junnyu/roformer_chinese_small''': 15_36, '''junnyu/roformer_chinese_base''': 15_36, '''junnyu/roformer_chinese_char_small''': 5_12, '''junnyu/roformer_chinese_char_base''': 5_12, '''junnyu/roformer_small_discriminator''': 1_28, '''junnyu/roformer_small_generator''': 1_28, } _UpperCAmelCase : Any = { '''junnyu/roformer_chinese_small''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_base''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_char_small''': {'''do_lower_case''': True}, '''junnyu/roformer_chinese_char_base''': {'''do_lower_case''': True}, '''junnyu/roformer_small_discriminator''': {'''do_lower_case''': True}, '''junnyu/roformer_small_generator''': {'''do_lower_case''': True}, } class lowercase_ ( _UpperCamelCase ): """simple docstring""" __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = PRETRAINED_INIT_CONFIGURATION __lowerCAmelCase = RoFormerTokenizer def __init__( self : Dict, UpperCamelCase__ : Dict=None, UpperCamelCase__ : List[Any]=None, UpperCamelCase__ : str=True, UpperCamelCase__ : Dict="[UNK]", UpperCamelCase__ : Dict="[SEP]", UpperCamelCase__ : Optional[Any]="[PAD]", UpperCamelCase__ : Union[str, Any]="[CLS]", UpperCamelCase__ : Any="[MASK]", UpperCamelCase__ : List[Any]=True, UpperCamelCase__ : str=None, **UpperCamelCase__ : str, ) -> Optional[int]: super().__init__( UpperCamelCase__, tokenizer_file=UpperCamelCase__, do_lower_case=UpperCamelCase__, unk_token=UpperCamelCase__, sep_token=UpperCamelCase__, pad_token=UpperCamelCase__, cls_token=UpperCamelCase__, mask_token=UpperCamelCase__, tokenize_chinese_chars=UpperCamelCase__, strip_accents=UpperCamelCase__, **UpperCamelCase__, ) _A = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get('lowercase', UpperCamelCase__ ) != do_lower_case or pre_tok_state.get('strip_accents', UpperCamelCase__ ) != strip_accents ): _A = getattr(UpperCamelCase__, pre_tok_state.pop('type' ) ) _A = do_lower_case _A = strip_accents _A = pre_tok_class(**UpperCamelCase__ ) _A = do_lower_case def __getstate__( self : Any ) -> List[str]: _A = self.__dict__.copy() _A = BertPreTokenizer() return state def __setstate__( self : List[str], UpperCamelCase__ : int ) -> int: _A = d _A = self.__dict__['_tokenizer'].get_vocab() _A = PreTokenizer.custom(JiebaPreTokenizer(UpperCamelCase__ ) ) def __UpperCAmelCase ( self : List[Any], UpperCamelCase__ : List[str], UpperCamelCase__ : int=None ) -> int: _A = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCAmelCase ( self : List[str], UpperCamelCase__ : List[int], UpperCamelCase__ : Optional[List[int]] = None ) -> List[int]: _A = [self.sep_token_id] _A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCAmelCase ( self : Tuple, UpperCamelCase__ : str, UpperCamelCase__ : Optional[str] = None ) -> Tuple[str]: _A = self._tokenizer.model.save(UpperCamelCase__, name=UpperCamelCase__ ) return tuple(UpperCamelCase__ ) def __UpperCAmelCase ( self : int, UpperCamelCase__ : int, UpperCamelCase__ : Dict=None, UpperCamelCase__ : Dict=None, UpperCamelCase__ : Tuple=False, **UpperCamelCase__ : Optional[Any], ) -> Any: _A = BertPreTokenizer() return super().save_pretrained(UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, **UpperCamelCase__ )
107
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if mass < 0: raise ValueError("""The mass of a body cannot be negative""" ) return 0.5 * mass * abs(__UpperCamelCase ) * abs(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
58
0
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase ( self : Optional[Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : List[Any] ) -> Tuple: """simple docstring""" return f"""gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy""" def lowerCamelCase ( self : List[Any] ) -> Any: """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() def lowerCamelCase ( self : int , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 4, 64, 64) , lowerCamelCase : Tuple=False ) -> Tuple: """simple docstring""" _UpperCAmelCase = jnp.bfloataa if fpaa else jnp.floataa _UpperCAmelCase = jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) , dtype=lowerCamelCase ) return image def lowerCamelCase ( self : Optional[int] , lowerCamelCase : Tuple=False , lowerCamelCase : str="CompVis/stable-diffusion-v1-4" ) -> Any: """simple docstring""" _UpperCAmelCase = jnp.bfloataa if fpaa else jnp.floataa _UpperCAmelCase = """bf16""" if fpaa else None _UpperCAmelCase , _UpperCAmelCase = FlaxUNetaDConditionModel.from_pretrained( lowerCamelCase , subfolder="""unet""" , dtype=lowerCamelCase , revision=lowerCamelCase ) return model, params def lowerCamelCase ( self : Tuple , lowerCamelCase : Any=0 , lowerCamelCase : Any=(4, 77, 768) , lowerCamelCase : Optional[Any]=False ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = jnp.bfloataa if fpaa else jnp.floataa _UpperCAmelCase = jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) , dtype=lowerCamelCase ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2323, -0.1304, 0.0813, -0.3093, -0.0919, -0.1571, -0.1125, -0.5806]], [17, 0.55, [-0.0831, -0.2443, 0.0901, -0.0919, 0.3396, 0.0103, -0.3743, 0.0701]], [8, 0.89, [-0.4863, 0.0859, 0.0875, -0.1658, 0.9199, -0.0114, 0.4839, 0.4639]], [3, 1000, [-0.5649, 0.2402, -0.5518, 0.1248, 1.1328, -0.2443, -0.0325, -1.0078]], # fmt: on ] ) def lowerCamelCase ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : str , lowerCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=lowerCamelCase ) _UpperCAmelCase = self.get_latents(lowerCamelCase , fpaa=lowerCamelCase ) _UpperCAmelCase = self.get_encoder_hidden_states(lowerCamelCase , fpaa=lowerCamelCase ) _UpperCAmelCase = model.apply( {"""params""": params} , lowerCamelCase , jnp.array(lowerCamelCase , dtype=jnp.intaa ) , encoder_hidden_states=lowerCamelCase , ).sample assert sample.shape == latents.shape _UpperCAmelCase = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) _UpperCAmelCase = jnp.array(lowerCamelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1514, 0.0807, 0.1624, 0.1016, -0.1896, 0.0263, 0.0677, 0.2310]], [17, 0.55, [0.1164, -0.0216, 0.0170, 0.1589, -0.3120, 0.1005, -0.0581, -0.1458]], [8, 0.89, [-0.1758, -0.0169, 0.1004, -0.1411, 0.1312, 0.1103, -0.1996, 0.2139]], [3, 1000, [0.1214, 0.0352, -0.0731, -0.1562, -0.0994, -0.0906, -0.2340, -0.0539]], # fmt: on ] ) def lowerCamelCase ( self : Any , lowerCamelCase : Any , lowerCamelCase : str , lowerCamelCase : List[str] ) -> List[str]: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=lowerCamelCase ) _UpperCAmelCase = self.get_latents(lowerCamelCase , shape=(4, 4, 96, 96) , fpaa=lowerCamelCase ) _UpperCAmelCase = self.get_encoder_hidden_states(lowerCamelCase , shape=(4, 77, 1024) , fpaa=lowerCamelCase ) _UpperCAmelCase = model.apply( {"""params""": params} , lowerCamelCase , jnp.array(lowerCamelCase , dtype=jnp.intaa ) , encoder_hidden_states=lowerCamelCase , ).sample assert sample.shape == latents.shape _UpperCAmelCase = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) _UpperCAmelCase = jnp.array(lowerCamelCase , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(lowerCamelCase , lowerCamelCase , atol=1E-2 )
108
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = StableDiffusionInpaintPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase = frozenset([] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=_lowercase , ) snake_case_ : Dict = PNDMScheduler(skip_prk_steps=_lowercase ) torch.manual_seed(0 ) snake_case_ : str = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) snake_case_ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) snake_case_ : Dict = CLIPTextModel(_lowercase ) snake_case_ : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case_ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self , _lowercase , _lowercase=0 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowercase ) ).to(_lowercase ) snake_case_ : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case_ : Tuple = Image.fromarray(np.uinta(_lowercase ) ).convert("""RGB""" ).resize((6_4, 6_4) ) snake_case_ : Any = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((6_4, 6_4) ) if str(_lowercase ).startswith("""mps""" ): snake_case_ : str = torch.manual_seed(_lowercase ) else: snake_case_ : List[str] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) snake_case_ : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case_ : List[str] = self.get_dummy_components() snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline(**_lowercase ) snake_case_ : Dict = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) snake_case_ : Optional[int] = self.get_dummy_inputs(_lowercase ) snake_case_ : List[str] = sd_pipe(**_lowercase ).images snake_case_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) snake_case_ : Optional[int] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained(_lowercase , safety_checker=_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : List[Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[str] = torch.manual_seed(0 ) snake_case_ : Dict = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ : str = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , torch_dtype=torch.floataa , safety_checker=_lowercase , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : Union[str, Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : Optional[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : str = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case_ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : int = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : List[str] = PNDMScheduler.from_pretrained(_lowercase , subfolder="""scheduler""" ) snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , safety_checker=_lowercase , scheduler=_lowercase , torch_dtype=torch.floataa , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case_ : Any = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , num_inference_steps=2 , output_type="""np""" , ) snake_case_ : Dict = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
58
0
'''simple docstring''' import argparse from pathlib import Path import requests import torch from PIL import Image from transformers import ( RobertaTokenizer, TrOCRConfig, TrOCRForCausalLM, TrOCRProcessor, VisionEncoderDecoderModel, ViTConfig, ViTImageProcessor, ViTModel, ) from transformers.utils import logging logging.set_verbosity_info() a = logging.get_logger(__name__) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' __SCREAMING_SNAKE_CASE = [] for i in range(encoder_config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f"""encoder.deit.blocks.{i}.norm1.weight""", f"""encoder.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""encoder.deit.blocks.{i}.norm1.bias""", f"""encoder.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (f"""encoder.deit.blocks.{i}.attn.proj.weight""", f"""encoder.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append( (f"""encoder.deit.blocks.{i}.attn.proj.bias""", f"""encoder.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append( (f"""encoder.deit.blocks.{i}.norm2.weight""", f"""encoder.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""encoder.deit.blocks.{i}.norm2.bias""", f"""encoder.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append( (f"""encoder.deit.blocks.{i}.mlp.fc1.weight""", f"""encoder.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append( (f"""encoder.deit.blocks.{i}.mlp.fc1.bias""", f"""encoder.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append( (f"""encoder.deit.blocks.{i}.mlp.fc2.weight""", f"""encoder.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""encoder.deit.blocks.{i}.mlp.fc2.bias""", f"""encoder.encoder.layer.{i}.output.dense.bias""") ) # cls token, position embeddings and patch embeddings of encoder rename_keys.extend( [ ("""encoder.deit.cls_token""", """encoder.embeddings.cls_token"""), ("""encoder.deit.pos_embed""", """encoder.embeddings.position_embeddings"""), ("""encoder.deit.patch_embed.proj.weight""", """encoder.embeddings.patch_embeddings.projection.weight"""), ("""encoder.deit.patch_embed.proj.bias""", """encoder.embeddings.patch_embeddings.projection.bias"""), ("""encoder.deit.norm.weight""", """encoder.layernorm.weight"""), ("""encoder.deit.norm.bias""", """encoder.layernorm.bias"""), ] ) return rename_keys def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' for i in range(encoder_config.num_hidden_layers ): # queries, keys and values (only weights, no biases) __SCREAMING_SNAKE_CASE = state_dict.pop(f"""encoder.deit.blocks.{i}.attn.qkv.weight""" ) __SCREAMING_SNAKE_CASE = in_proj_weight[ : encoder_config.hidden_size, : ] __SCREAMING_SNAKE_CASE = in_proj_weight[ encoder_config.hidden_size : encoder_config.hidden_size * 2, : ] __SCREAMING_SNAKE_CASE = in_proj_weight[ -encoder_config.hidden_size :, : ] def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' __SCREAMING_SNAKE_CASE = dct.pop(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = val def __magic_name__ ( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if "handwritten" in checkpoint_url: __SCREAMING_SNAKE_CASE = """https://fki.tic.heia-fr.ch/static/img/a01-122-02-00.jpg""" # industry # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-12.jpg" # have # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-10.jpg" # let # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02.jpg" # # url = "https://fki.tic.heia-fr.ch/static/img/a01-122.jpg" elif "printed" in checkpoint_url or "stage1" in checkpoint_url: __SCREAMING_SNAKE_CASE = """https://www.researchgate.net/profile/Dinh-Sang/publication/338099565/figure/fig8/AS:840413229350922@1577381536857/An-receipt-example-in-the-SROIE-2019-dataset_Q640.jpg""" __SCREAMING_SNAKE_CASE = Image.open(requests.get(__UpperCAmelCase , stream=__UpperCAmelCase ).raw ).convert("""RGB""" ) return im @torch.no_grad() def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> int: '''simple docstring''' __SCREAMING_SNAKE_CASE = ViTConfig(image_size=384 , qkv_bias=__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = TrOCRConfig() # size of the architecture if "base" in checkpoint_url: __SCREAMING_SNAKE_CASE = 768 elif "large" in checkpoint_url: # use ViT-large encoder __SCREAMING_SNAKE_CASE = 1024 __SCREAMING_SNAKE_CASE = 4096 __SCREAMING_SNAKE_CASE = 24 __SCREAMING_SNAKE_CASE = 16 __SCREAMING_SNAKE_CASE = 1024 else: raise ValueError("""Should either find 'base' or 'large' in checkpoint URL""" ) # the large-printed + stage1 checkpoints uses sinusoidal position embeddings, no layernorm afterwards if "large-printed" in checkpoint_url or "stage1" in checkpoint_url: __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = """relu""" __SCREAMING_SNAKE_CASE = 1024 __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False # load HuggingFace model __SCREAMING_SNAKE_CASE = ViTModel(__UpperCAmelCase , add_pooling_layer=__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = TrOCRForCausalLM(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = VisionEncoderDecoderModel(encoder=__UpperCAmelCase , decoder=__UpperCAmelCase ) model.eval() # load state_dict of original model, rename some keys __SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(__UpperCAmelCase , map_location="""cpu""" , check_hash=__UpperCAmelCase )["""model"""] __SCREAMING_SNAKE_CASE = create_rename_keys(__UpperCAmelCase , __UpperCAmelCase ) for src, dest in rename_keys: rename_key(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) read_in_q_k_v(__UpperCAmelCase , __UpperCAmelCase ) # remove parameters we don't need del state_dict["encoder.deit.head.weight"] del state_dict["encoder.deit.head.bias"] del state_dict["decoder.version"] # add prefix to decoder keys for key, val in state_dict.copy().items(): __SCREAMING_SNAKE_CASE = state_dict.pop(__UpperCAmelCase ) if key.startswith("""decoder""" ) and "output_projection" not in key: __SCREAMING_SNAKE_CASE = val else: __SCREAMING_SNAKE_CASE = val # load state dict model.load_state_dict(__UpperCAmelCase ) # Check outputs on an image __SCREAMING_SNAKE_CASE = ViTImageProcessor(size=encoder_config.image_size ) __SCREAMING_SNAKE_CASE = RobertaTokenizer.from_pretrained("""roberta-large""" ) __SCREAMING_SNAKE_CASE = TrOCRProcessor(__UpperCAmelCase , __UpperCAmelCase ) __SCREAMING_SNAKE_CASE = processor(images=prepare_img(__UpperCAmelCase ) , return_tensors="""pt""" ).pixel_values # verify logits __SCREAMING_SNAKE_CASE = torch.tensor([[model.config.decoder.decoder_start_token_id]] ) __SCREAMING_SNAKE_CASE = model(pixel_values=__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = outputs.logits __SCREAMING_SNAKE_CASE = torch.Size([1, 1, 50265] ) if "trocr-base-handwritten" in checkpoint_url: __SCREAMING_SNAKE_CASE = torch.tensor( [-1.4_5_0_2, -4.6_6_8_3, -0.5_3_4_7, -2.9_2_9_1, 9.1_4_3_5, -3.0_5_7_1, 8.9_7_6_4, 1.7_5_6_0, 8.7_3_5_8, -1.5_3_1_1] ) elif "trocr-large-handwritten" in checkpoint_url: __SCREAMING_SNAKE_CASE = torch.tensor( [-2.6_4_3_7, -1.3_1_2_9, -2.2_5_9_6, -5.3_4_5_5, 6.3_5_3_9, 1.7_6_0_4, 5.4_9_9_1, 1.4_7_0_2, 5.6_1_1_3, 2.0_1_7_0] ) elif "trocr-base-printed" in checkpoint_url: __SCREAMING_SNAKE_CASE = torch.tensor( [-5.6_8_1_6, -5.8_3_8_8, 1.1_3_9_8, -6.9_0_3_4, 6.8_5_0_5, -2.4_3_9_3, 1.2_2_8_4, -1.0_2_3_2, -1.9_6_6_1, -3.9_2_1_0] ) elif "trocr-large-printed" in checkpoint_url: __SCREAMING_SNAKE_CASE = torch.tensor( [-6.0_1_6_2, -7.0_9_5_9, 4.4_1_5_5, -5.1_0_6_3, 7.0_4_6_8, -3.1_6_3_1, 2.6_4_6_6, -0.3_0_8_1, -0.8_1_0_6, -1.7_5_3_5] ) if "stage1" not in checkpoint_url: assert logits.shape == expected_shape, "Shape of logits not as expected" assert torch.allclose(logits[0, 0, :10] , __UpperCAmelCase , atol=1e-3 ), "First elements of logits not as expected" Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(__UpperCAmelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(__UpperCAmelCase ) if __name__ == "__main__": a = argparse.ArgumentParser() parser.add_argument( "--checkpoint_url", default="https://layoutlm.blob.core.windows.net/trocr/model_zoo/fairseq/trocr-base-handwritten.pt", type=str, help="URL to the original PyTorch checkpoint (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) a = parser.parse_args() convert_tr_ocr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
109
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : str , __UpperCamelCase : Any ): '''simple docstring''' snake_case_ : Optional[Any] = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] snake_case_ : Optional[int] = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } snake_case_ : Optional[Any] = F'{src_lang}-{tgt_lang}' snake_case_ : Dict = F'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) snake_case_ : List[str] = os.path.join(__UpperCamelCase , """README.md""" ) print(F'Generating {path}' ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(__UpperCamelCase ) # make sure we are under the root of the project __lowerCAmelCase : str = Path(__file__).resolve().parent.parent.parent __lowerCAmelCase : Optional[int] = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[int] = model_name.split('''-''') __lowerCAmelCase : Optional[int] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
58
0
"""simple docstring""" import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'vocab_file': 'spiece.model'} UpperCamelCase__ = { 'vocab_file': { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model', 'google/bigbird-roberta-large': ( 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model' ), 'google/bigbird-base-trivia-itc': ( 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model' ), } } UpperCamelCase__ = { 'google/bigbird-roberta-base': 40_96, 'google/bigbird-roberta-large': 40_96, 'google/bigbird-base-trivia-itc': 40_96, } class a ( lowercase ): UpperCamelCase : Dict = VOCAB_FILES_NAMES UpperCamelCase : int = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : List[str] = ["""input_ids""", """attention_mask"""] UpperCamelCase : List[int] = [] def __init__( self , UpperCamelCase_ , UpperCamelCase_="<unk>" , UpperCamelCase_="<s>" , UpperCamelCase_="</s>" , UpperCamelCase_="<pad>" , UpperCamelCase_="[SEP]" , UpperCamelCase_="[MASK]" , UpperCamelCase_="[CLS]" , UpperCamelCase_ = None , **UpperCamelCase_ , ): UpperCAmelCase__ : int = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else bos_token UpperCAmelCase__ : int = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else eos_token UpperCAmelCase__ : List[Any] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else unk_token UpperCAmelCase__ : Optional[int] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else pad_token UpperCAmelCase__ : Optional[Any] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else cls_token UpperCAmelCase__ : str = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else mask_token UpperCAmelCase__ : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) UpperCAmelCase__ : Any = vocab_file UpperCAmelCase__ : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase_ ) @property def __snake_case ( self ): return self.sp_model.get_piece_size() def __snake_case ( self ): UpperCAmelCase__ : Optional[Any] = {self.convert_ids_to_tokens(UpperCamelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): UpperCAmelCase__ : str = self.__dict__.copy() UpperCAmelCase__ : Dict = None return state def __setstate__( self , UpperCamelCase_ ): UpperCAmelCase__ : List[Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): UpperCAmelCase__ : Any = {} UpperCAmelCase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __snake_case ( self , UpperCamelCase_ ): return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_ ) def __snake_case ( self , UpperCamelCase_ ): return self.sp_model.piece_to_id(UpperCamelCase_ ) def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : int = self.sp_model.IdToPiece(UpperCamelCase_ ) return token def __snake_case ( self , UpperCamelCase_ ): UpperCAmelCase__ : List[str] = [] UpperCAmelCase__ : Optional[Any] = '' UpperCAmelCase__ : str = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(UpperCamelCase_ ) + token UpperCAmelCase__ : Optional[Any] = True UpperCAmelCase__ : str = [] else: current_sub_tokens.append(UpperCamelCase_ ) UpperCAmelCase__ : Tuple = False out_string += self.sp_model.decode(UpperCamelCase_ ) return out_string.strip() def __snake_case ( self , UpperCamelCase_ , UpperCamelCase_ = False , UpperCamelCase_ = None , UpperCamelCase_ = True , **UpperCamelCase_ , ): UpperCAmelCase__ : Any = kwargs.pop('use_source_tokenizer' , UpperCamelCase_ ) UpperCAmelCase__ : List[Any] = self.convert_ids_to_tokens(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 UpperCAmelCase__ : List[str] = [] UpperCAmelCase__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCamelCase_ ) ) UpperCAmelCase__ : Dict = [] sub_texts.append(UpperCamelCase_ ) else: current_sub_text.append(UpperCamelCase_ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCamelCase_ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: UpperCAmelCase__ : List[str] = re.sub(R' (\[(MASK|SEP)\])' , R'\1' , ' '.join(UpperCamelCase_ ) ) else: UpperCAmelCase__ : str = ''.join(UpperCamelCase_ ) UpperCAmelCase__ : Union[str, Any] = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: UpperCAmelCase__ : str = self.clean_up_tokenization(UpperCamelCase_ ) return clean_text else: return text def __snake_case ( self , UpperCamelCase_ , UpperCamelCase_ = None ): if not os.path.isdir(UpperCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase_ , 'wb' ) as fi: UpperCAmelCase__ : str = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_ ) return (out_vocab_file,) def __snake_case ( self , UpperCamelCase_ , UpperCamelCase_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] UpperCAmelCase__ : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def __snake_case ( self , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase_ )) + [1] return [1] + ([0] * len(UpperCamelCase_ )) + [1] + ([0] * len(UpperCamelCase_ )) + [1] def __snake_case ( self , UpperCamelCase_ , UpperCamelCase_ = None ): UpperCAmelCase__ : str = [self.sep_token_id] UpperCAmelCase__ : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
110
"""simple docstring""" __lowerCAmelCase : Tuple = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowerCAmelCase : Tuple = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __lowerCAmelCase : Any = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
58
0
import datasets from .evaluate import evaluate _lowerCAmelCase : Tuple ='''\ @article{hendrycks2021cuad, title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review}, author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball}, journal={arXiv preprint arXiv:2103.06268}, year={2021} } ''' _lowerCAmelCase : Union[str, Any] =''' This metric wrap the official scoring script for version 1 of the Contract Understanding Atticus Dataset (CUAD). Contract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510 commercial legal contracts that have been manually labeled to identify 41 categories of important clauses that lawyers look for when reviewing contracts in connection with corporate transactions. ''' _lowerCAmelCase : Optional[Any] =''' Computes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall). Args: predictions: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair as given in the references (see below) - \'prediction_text\': list of possible texts for the answer, as a list of strings depending on a threshold on the confidence probability of each prediction. references: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair (see above), - \'answers\': a Dict in the CUAD dataset format { \'text\': list of possible texts for the answer, as a list of strings \'answer_start\': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: \'exact_match\': Exact match (the normalized answer exactly match the gold answer) \'f1\': The F-score of predicted tokens versus the gold answer \'aupr\': Area Under the Precision-Recall curve \'prec_at_80_recall\': Precision at 80% recall \'prec_at_90_recall\': Precision at 90% recall Examples: >>> predictions = [{\'prediction_text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\'], \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}] >>> references = [{\'answers\': {\'answer_start\': [143, 49], \'text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\']}, \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}] >>> cuad_metric = datasets.load_metric("cuad") >>> results = cuad_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 100.0, \'f1\': 100.0, \'aupr\': 0.0, \'prec_at_80_recall\': 1.0, \'prec_at_90_recall\': 1.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): '''simple docstring''' def _UpperCAmelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": { "id": datasets.Value("string" ), "prediction_text": datasets.features.Sequence(datasets.Value("string" ) ), }, "references": { "id": datasets.Value("string" ), "answers": datasets.features.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), }, } ) , codebase_urls=["https://www.atticusprojectai.org/cuad"] , reference_urls=["https://www.atticusprojectai.org/cuad"] , ) def _UpperCAmelCase ( self , lowerCamelCase__ , lowerCamelCase__ ): UpperCAmelCase__: Union[str, Any] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} UpperCAmelCase__: Union[str, Any] = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] UpperCAmelCase__: int = evaluate(dataset=_lowercase , predictions=_lowercase ) return score
113
"""simple docstring""" from jiwer import compute_measures import datasets __lowerCAmelCase : Tuple = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __lowerCAmelCase : Union[str, Any] = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' __lowerCAmelCase : Optional[int] = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", ] , ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=False ) -> Optional[Any]: '''simple docstring''' if concatenate_texts: return compute_measures(_lowercase , _lowercase )["wer"] else: snake_case_ : List[str] = 0 snake_case_ : Optional[int] = 0 for prediction, reference in zip(_lowercase , _lowercase ): snake_case_ : Optional[Any] = compute_measures(_lowercase , _lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
58
0
"""simple docstring""" import numpy # List of input, output pairs __UpperCamelCase : List[str] = ( ((5, 2, 3), 1_5), ((6, 5, 9), 2_5), ((1_1, 1_2, 1_3), 4_1), ((1, 1, 1), 8), ((1_1, 1_2, 1_3), 4_1), ) __UpperCamelCase : List[Any] = (((5_1_5, 2_2, 1_3), 5_5_5), ((6_1, 3_5, 4_9), 1_5_0)) __UpperCamelCase : Union[str, Any] = [2, 4, 1, 5] __UpperCamelCase : Dict = len(train_data) __UpperCamelCase : int = 0.0_0_9 def __SCREAMING_SNAKE_CASE ( A_ , A_="train" ): return calculate_hypothesis_value(__UpperCamelCase , __UpperCamelCase ) - output( __UpperCamelCase , __UpperCamelCase ) def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Optional[Any] = 0 for i in range(len(__UpperCamelCase ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def __SCREAMING_SNAKE_CASE ( A_ , A_ ): if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def __SCREAMING_SNAKE_CASE ( A_ , A_ ): if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def __SCREAMING_SNAKE_CASE ( A_ , A_=m ): lowerCAmelCase__ : Union[str, Any] = 0 for i in range(__UpperCamelCase ): if index == -1: summation_value += _error(__UpperCamelCase ) else: summation_value += _error(__UpperCamelCase ) * train_data[i][0][index] return summation_value def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Optional[Any] = summation_of_cost_derivative(__UpperCamelCase , __UpperCamelCase ) / m return cost_derivative_value def __SCREAMING_SNAKE_CASE ( ): global parameter_vector # Tune these values to set a tolerance value for predicted output lowerCAmelCase__ : Optional[int] = 0.000_002 lowerCAmelCase__ : Optional[Any] = 0 lowerCAmelCase__ : Tuple = 0 while True: j += 1 lowerCAmelCase__ : str = [0, 0, 0, 0] for i in range(0 , len(__UpperCamelCase ) ): lowerCAmelCase__ : List[str] = get_cost_derivative(i - 1 ) lowerCAmelCase__ : List[str] = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( __UpperCamelCase , __UpperCamelCase , atol=__UpperCamelCase , rtol=__UpperCamelCase , ): break lowerCAmelCase__ : List[str] = temp_parameter_vector print(('''Number of iterations:''', j) ) def __SCREAMING_SNAKE_CASE ( ): for i in range(len(__UpperCamelCase ) ): print(('''Actual output value:''', output(__UpperCamelCase , '''test''' )) ) print(('''Hypothesis output:''', calculate_hypothesis_value(__UpperCamelCase , '''test''' )) ) if __name__ == "__main__": run_gradient_descent() print('''\nTesting gradient descent for a linear hypothesis function.\n''') test_gradient_descent()
450
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1_3 , _lowercase=3 , _lowercase=2_2_4 , _lowercase=3_0 , _lowercase=4_0_0 , _lowercase=True , _lowercase=None , _lowercase=True , _lowercase=[0.5, 0.5, 0.5] , _lowercase=[0.5, 0.5, 0.5] , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : Optional[Any] = batch_size snake_case_ : Dict = num_channels snake_case_ : Optional[Any] = image_size snake_case_ : Optional[Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[int] = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : Dict = image_std def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ViTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = EfficientFormerImageProcessorTester(self ) @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.image_proc_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , """image_mean""" ) ) self.assertTrue(hasattr(_lowercase , """image_std""" ) ) self.assertTrue(hasattr(_lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowercase , """do_resize""" ) ) self.assertTrue(hasattr(_lowercase , """size""" ) ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' pass def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input snake_case_ : Any = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Optional[Any] = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Dict = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input snake_case_ : int = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : int = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input snake_case_ : Dict = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Tuple = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
58
0
import os from pathlib import Path def lowerCamelCase ( a_ , a_ , a_ ) -> List[str]: lowerCAmelCase_ = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] lowerCAmelCase_ = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } lowerCAmelCase_ = F'''{src_lang}-{tgt_lang}''' lowerCAmelCase_ = F'''\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n''' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) lowerCAmelCase_ = os.path.join(__UpperCamelCase , 'README.md' ) print(F'''Generating {path}''' ) with open(__UpperCamelCase , 'w' , encoding='utf-8' ) as f: f.write(__UpperCamelCase ) # make sure we are under the root of the project lowerCamelCase_ = Path(__file__).resolve().parent.parent.parent lowerCamelCase_ = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowerCamelCase_ = model_name.split("""-""") lowerCamelCase_ = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
318
"""simple docstring""" from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCAmelCase : int = TypeVar('''KT''') __lowerCAmelCase : Union[str, Any] = TypeVar('''VT''') class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = "root" , _lowercase = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = key snake_case_ : Tuple = value snake_case_ : list[Node[KT, VT]] = [] def __repr__( self ) -> str: '''simple docstring''' return f'Node({self.key}: {self.value})' @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return len(self.forward ) class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = 0.5 , _lowercase = 1_6 ) -> int: '''simple docstring''' snake_case_ : Node[KT, VT] = Node[KT, VT]() snake_case_ : Union[str, Any] = 0 snake_case_ : Optional[int] = p snake_case_ : Any = max_level def __str__( self ) -> str: '''simple docstring''' snake_case_ : str = list(self ) if len(_lowercase ) == 0: return f'SkipList(level={self.level})' snake_case_ : List[Any] = max((len(str(_lowercase ) ) for item in items) , default=4 ) snake_case_ : str = max(_lowercase , 4 ) + 4 snake_case_ : Union[str, Any] = self.head snake_case_ : Dict = [] snake_case_ : List[str] = node.forward.copy() lines.append(f'[{node.key}]'.ljust(_lowercase , """-""" ) + """* """ * len(_lowercase ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) while len(node.forward ) != 0: snake_case_ : Optional[Any] = node.forward[0] lines.append( f'[{node.key}]'.ljust(_lowercase , """-""" ) + """ """.join(str(n.key ) if n.key == node.key else """|""" for n in forwards ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) snake_case_ : List[str] = node.forward lines.append("""None""".ljust(_lowercase ) + """* """ * len(_lowercase ) ) return f'SkipList(level={self.level})\n' + "\n".join(_lowercase ) def __iter__( self ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.head while len(node.forward ) != 0: yield node.forward[0].key snake_case_ : Dict = node.forward[0] def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[int] = 1 while random() < self.p and level < self.max_level: level += 1 return level def UpperCAmelCase__ ( self , _lowercase ) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: snake_case_ : List[Any] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(_lowercase ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: for i, update_node in enumerate(_lowercase ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: snake_case_ : List[str] = node.forward[i] else: snake_case_ : Tuple = update_node.forward[:i] def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> str: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: snake_case_ : List[Any] = value else: snake_case_ : Optional[int] = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , _lowercase ): update_vector.append(self.head ) snake_case_ : Any = level snake_case_ : Optional[int] = Node(_lowercase , _lowercase ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(_lowercase ) else: snake_case_ : Optional[Any] = new_node def UpperCAmelCase__ ( self , _lowercase ) -> VT | None: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: return node.value return None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : List[str] = SkipList() skip_list.insert("""Key1""" , 3 ) skip_list.insert("""Key2""" , 1_2 ) skip_list.insert("""Key3""" , 4_1 ) skip_list.insert("""Key4""" , -1_9 ) snake_case_ : Optional[int] = skip_list.head snake_case_ : List[Any] = {} while node.level != 0: snake_case_ : List[str] = node.forward[0] snake_case_ : Union[str, Any] = node.value assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 1_2 assert all_values["Key3"] == 4_1 assert all_values["Key4"] == -1_9 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_0 ) skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""Key5""" , 7 ) skip_list.insert("""Key7""" , 1_0 ) skip_list.insert("""Key10""" , 5 ) skip_list.insert("""Key7""" , 7 ) skip_list.insert("""Key5""" , 5 ) skip_list.insert("""Key10""" , 1_0 ) snake_case_ : str = skip_list.head snake_case_ : str = {} while node.level != 0: snake_case_ : Optional[Any] = node.forward[0] snake_case_ : int = node.value if len(__UpperCamelCase ) != 4: print() assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 1_2 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 1_0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : str = SkipList() assert skip_list.find("""Some key""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[Any] = SkipList() skip_list.insert("""Key2""" , 2_0 ) assert skip_list.find("""Key2""" ) == 2_0 skip_list.insert("""Some Key""" , 1_0 ) skip_list.insert("""Key2""" , 8 ) skip_list.insert("""V""" , 1_3 ) assert skip_list.find("""Y""" ) is None assert skip_list.find("""Key2""" ) == 8 assert skip_list.find("""Some Key""" ) == 1_0 assert skip_list.find("""V""" ) == 1_3 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Any = SkipList() skip_list.delete("""Some key""" ) assert len(skip_list.head.forward ) == 0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Tuple = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) == 1_4 assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""X""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key1""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Union[str, Any] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4_2 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""X""" ) def traverse_keys(__UpperCamelCase : str ): yield node.key for forward_node in node.forward: yield from traverse_keys(__UpperCamelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __lowerCAmelCase ( ): '''simple docstring''' def is_sorted(__UpperCamelCase : List[Any] ): return all(next_item >= item for item, next_item in zip(__UpperCamelCase , lst[1:] ) ) snake_case_ : str = SkipList() for i in range(1_0 ): skip_list.insert(__UpperCamelCase , __UpperCamelCase ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.insert(-1_2 , -1_2 ) skip_list.insert(7_7 , 7_7 ) assert is_sorted(list(__UpperCamelCase ) ) def __lowerCAmelCase ( ): '''simple docstring''' for _ in range(1_0_0 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = SkipList() skip_list.insert(2 , """2""" ) skip_list.insert(4 , """4""" ) skip_list.insert(6 , """4""" ) skip_list.insert(4 , """5""" ) skip_list.insert(8 , """4""" ) skip_list.insert(9 , """4""" ) skip_list.delete(4 ) print(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
58
0
"""simple docstring""" from queue import PriorityQueue from typing import Any import numpy as np def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, __snake_case, ) -> Optional[Any]: """simple docstring""" for nxt, d in graph[v]: if nxt in visited_forward: continue _UpperCamelCase = cst_fwd.get(__UpperCamelCase, np.inf ) _UpperCamelCase = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) _UpperCamelCase = new_cost_f _UpperCamelCase = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: _UpperCamelCase = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = -1 _UpperCamelCase = set() _UpperCamelCase = set() _UpperCamelCase = {source: 0} _UpperCamelCase = {destination: 0} _UpperCamelCase = {source: None} _UpperCamelCase = {destination: None} _UpperCamelCase = PriorityQueue() _UpperCamelCase = PriorityQueue() _UpperCamelCase = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): _UpperCamelCase = queue_forward.get() visited_forward.add(__UpperCamelCase ) _UpperCamelCase = queue_backward.get() visited_backward.add(__UpperCamelCase ) _UpperCamelCase = pass_and_relaxation( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, ) _UpperCamelCase = pass_and_relaxation( __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, __UpperCamelCase, ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: _UpperCamelCase = shortest_distance return shortest_path_distance _a = { '''B''': [['''C''', 1]], '''C''': [['''D''', 1]], '''D''': [['''F''', 1]], '''E''': [['''B''', 1], ['''G''', 2]], '''F''': [], '''G''': [['''F''', 1]], } _a = { '''B''': [['''E''', 1]], '''C''': [['''B''', 1]], '''D''': [['''C''', 1]], '''F''': [['''D''', 1], ['''G''', 1]], '''E''': [[None, np.inf]], '''G''': [['''E''', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
19
"""simple docstring""" import argparse import os import re import packaging.version __lowerCAmelCase : Optional[Any] = '''examples/''' __lowerCAmelCase : Union[str, Any] = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __lowerCAmelCase : Union[str, Any] = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } __lowerCAmelCase : List[Any] = '''README.md''' def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : Dict ): '''simple docstring''' with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : Any = f.read() snake_case_ , snake_case_ : Optional[int] = REPLACE_PATTERNS[pattern] snake_case_ : Union[str, Any] = replace.replace("""VERSION""" , __UpperCamelCase ) snake_case_ : List[Any] = re_pattern.sub(__UpperCamelCase , __UpperCamelCase ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' for folder, directories, fnames in os.walk(__UpperCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(__UpperCamelCase , __UpperCamelCase ) , __UpperCamelCase , pattern="""examples""" ) def __lowerCAmelCase ( __UpperCamelCase : List[str] , __UpperCamelCase : int=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not patch: update_version_in_examples(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = """🤗 Transformers currently provides the following architectures""" snake_case_ : Union[str, Any] = """1. Want to contribute a new model?""" with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : str = f.readlines() # Find the start of the list. snake_case_ : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 snake_case_ : Optional[int] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): snake_case_ : Any = lines[index].replace( """https://huggingface.co/docs/diffusers/main/model_doc""" , """https://huggingface.co/docs/diffusers/model_doc""" , ) index += 1 with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] , """r""" ) as f: snake_case_ : Any = f.read() snake_case_ : Tuple = REPLACE_PATTERNS["""init"""][0].search(__UpperCamelCase ).groups()[0] return packaging.version.parse(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : str=False ): '''simple docstring''' snake_case_ : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: snake_case_ : str = default_version.base_version elif patch: snake_case_ : str = F'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: snake_case_ : str = F'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. snake_case_ : int = input(F'Which version are you releasing? [{default_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Optional[int] = default_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase , patch=__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = get_version() snake_case_ : str = F'{current_version.major}.{current_version.minor + 1}.0.dev0' snake_case_ : Tuple = current_version.base_version # Check with the user we got that right. snake_case_ : Optional[int] = input(F'Which version are we developing now? [{dev_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Dict = dev_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": __lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __lowerCAmelCase : str = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
58
0
'''simple docstring''' from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class lowerCAmelCase__ : """simple docstring""" lowerCAmelCase__ = 42 lowerCAmelCase__ = None lowerCAmelCase__ = None UpperCAmelCase : Dict = namedtuple('CoinsDistribResult', 'moves excess') def a__ ( a__ ): """simple docstring""" if root is None: return 0 # Validation def count_nodes(a__ ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(a__ ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(__UpperCamelCase ) != count_coins(__UpperCamelCase ): raise ValueError("""The nodes number should be same as the number of coins""" ) # Main calculation def get_distrib(a__ ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) __SCREAMING_SNAKE_CASE = get_distrib(node.left ) __SCREAMING_SNAKE_CASE = get_distrib(node.right ) __SCREAMING_SNAKE_CASE = 1 - left_distrib_excess __SCREAMING_SNAKE_CASE = 1 - right_distrib_excess __SCREAMING_SNAKE_CASE = ( left_distrib_moves + right_distrib_moves + abs(__UpperCamelCase ) + abs(__UpperCamelCase ) ) __SCREAMING_SNAKE_CASE = node.data - coins_to_left - coins_to_right return CoinsDistribResult(__UpperCamelCase , __UpperCamelCase ) return get_distrib(__UpperCamelCase )[0] if __name__ == "__main__": import doctest doctest.testmod()
627
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
58
0
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def a (self : List[Any] ): """simple docstring""" __snake_case = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) __snake_case = Vector() def a (self : Optional[int] ): """simple docstring""" __snake_case = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(_lowercase ) , '''(0,0,0,0,0,1)''' ) def a (self : List[Any] ): """simple docstring""" __snake_case = Vector([1, 2, 3, 4] ) self.assertEqual(len(_lowercase ) , 4 ) def a (self : Union[str, Any] ): """simple docstring""" __snake_case = Vector([1, 2] ) __snake_case = Vector([1, 2, 3, 4, 5] ) __snake_case = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) __snake_case = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.2_3_6 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.4_1_6 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.6_1_6 , 3 ) def a (self : int ): """simple docstring""" __snake_case = Vector([1, 2, 3] ) __snake_case = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def a (self : Dict ): """simple docstring""" __snake_case = Vector([1, 2, 3] ) __snake_case = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def a (self : List[Any] ): """simple docstring""" __snake_case = Vector([1, 2, 3] ) __snake_case = Vector([2, -1, 4] ) # for test of dot product __snake_case = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , '''(3.0,6.0,9.0)''' ) self.assertEqual((a * b) , 0 ) def a (self : Any ): """simple docstring""" self.assertEqual(str(zero_vector(10 ) ).count('''0''' ) , 10 ) def a (self : List[str] ): """simple docstring""" self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , '''(0,1,0)''' ) def a (self : Dict ): """simple docstring""" __snake_case = Vector([1, 2, 3] ) __snake_case = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , _lowercase , _lowercase ) ) , '''(3,4,7)''' ) def a (self : Optional[int] ): """simple docstring""" __snake_case = Vector([1, 0, 0, 0, 0, 0] ) __snake_case = x.copy() self.assertEqual(str(_lowercase ) , str(_lowercase ) ) def a (self : int ): """simple docstring""" __snake_case = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(_lowercase ) , '''(0,1,0)''' ) def a (self : List[Any] ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual('''|1,2,3|\n|2,4,5|\n|6,7,8|\n''' , str(_lowercase ) ) def a (self : Dict ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __snake_case = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(_lowercase , _lowercase ) ) def a (self : str ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __snake_case = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(_lowercase , _lowercase ) ) def a (self : str ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def a (self : List[str] ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) __snake_case = Vector([1, 2, 3] ) self.assertEqual('''(14,32,50)''' , str(a * x ) ) self.assertEqual('''|2,4,6|\n|8,10,12|\n|14,16,18|\n''' , str(a * 2 ) ) def a (self : Any ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual('''|1,2,5|\n|2,4,5|\n|6,7,8|\n''' , str(_lowercase ) ) def a (self : Dict ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.0_1 ) def a (self : Optional[int] ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __snake_case = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|2,4,10|\n|4,8,10|\n|12,14,18|\n''' , str(a + b ) ) def a (self : List[Any] ): """simple docstring""" __snake_case = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) __snake_case = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual('''|0,0,-4|\n|0,0,0|\n|0,0,-2|\n''' , str(a - b ) ) def a (self : Any ): """simple docstring""" self.assertEqual( '''|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n''' , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
592
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) snake_case_ : str = precision snake_case_ : Any = ceil(precision / 1_4 ) snake_case_ : Dict = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : Optional[Any] = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : Optional[int] = Decimal(__UpperCamelCase ) for k in range(1 , __UpperCamelCase ): snake_case_ : Any = factorial(6 * k ) // (factorial(3 * k ) * factorial(__UpperCamelCase ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __lowerCAmelCase : int = 50 print(F'''The first {n} digits of pi is: {pi(n)}''')
58
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowercase__ = logging.get_logger(__name__) lowercase__ = { '''shi-labs/dinat-mini-in1k-224''': '''https://huggingface.co/shi-labs/dinat-mini-in1k-224/resolve/main/config.json''', # See all Dinat models at https://huggingface.co/models?filter=dinat } class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _lowerCAmelCase = "dinat" _lowerCAmelCase = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__(self , _lowercase=4 , _lowercase=3 , _lowercase=64 , _lowercase=[3, 4, 6, 5] , _lowercase=[2, 4, 8, 16] , _lowercase=7 , _lowercase=[[1, 8, 1], [1, 4, 1, 4], [1, 2, 1, 2, 1, 2], [1, 1, 1, 1, 1]] , _lowercase=3.0 , _lowercase=True , _lowercase=0.0 , _lowercase=0.0 , _lowercase=0.1 , _lowercase="gelu" , _lowercase=0.02 , _lowercase=1e-5 , _lowercase=0.0 , _lowercase=None , _lowercase=None , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : Dict = patch_size __a : Union[str, Any] = num_channels __a : Optional[Any] = embed_dim __a : List[Any] = depths __a : Any = len(_lowercase ) __a : str = num_heads __a : Dict = kernel_size __a : Optional[int] = dilations __a : List[Any] = mlp_ratio __a : List[str] = qkv_bias __a : List[str] = hidden_dropout_prob __a : Optional[Any] = attention_probs_dropout_prob __a : int = drop_path_rate __a : int = hidden_act __a : str = layer_norm_eps __a : Union[str, Any] = initializer_range # we set the hidden_size attribute in order to make Dinat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __a : str = int(embed_dim * 2 ** (len(_lowercase ) - 1) ) __a : Optional[Any] = layer_scale_init_value __a : Optional[int] = ["""stem"""] + [F'''stage{idx}''' for idx in range(1 , len(_lowercase ) + 1 )] __a : Union[str, Any] = get_aligned_output_features_output_indices( out_features=_lowercase , out_indices=_lowercase , stage_names=self.stage_names )
581
"""simple docstring""" import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCAmelCase ( __UpperCamelCase : Optional[int] ): '''simple docstring''' snake_case_ : Any = torch.exp(__UpperCamelCase ) snake_case_ : Optional[int] = torch.sum(__UpperCamelCase , dim=1 ) # sum of exp(x_i) snake_case_ : str = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(__UpperCamelCase ) - B / A class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> int: '''simple docstring''' super().__init__() snake_case_ : Tuple = config.output_attentions snake_case_ : str = config.output_hidden_states snake_case_ : List[str] = nn.ModuleList([BertLayer(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Tuple = nn.ModuleList([BertHighway(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Any = [-1 for _ in range(config.num_hidden_layers )] def UpperCAmelCase__ ( self , _lowercase ) -> Tuple: '''simple docstring''' if (type(_lowercase ) is float) or (type(_lowercase ) is int): for i in range(len(self.early_exit_entropy ) ): snake_case_ : Dict = x else: snake_case_ : Union[str, Any] = x def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : int = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Any: '''simple docstring''' snake_case_ : str = () snake_case_ : str = () snake_case_ : List[str] = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: snake_case_ : int = all_hidden_states + (hidden_states,) snake_case_ : Any = layer_module( _lowercase , _lowercase , head_mask[i] , _lowercase , _lowercase ) snake_case_ : Dict = layer_outputs[0] if self.output_attentions: snake_case_ : str = all_attentions + (layer_outputs[1],) snake_case_ : Optional[int] = (hidden_states,) if self.output_hidden_states: snake_case_ : Tuple = current_outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : int = current_outputs + (all_attentions,) snake_case_ : Optional[Any] = self.highway[i](_lowercase ) # logits, pooled_output if not self.training: snake_case_ : Tuple = highway_exit[0] snake_case_ : List[str] = entropy(_lowercase ) snake_case_ : str = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy snake_case_ : Union[str, Any] = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: snake_case_ : List[Any] = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(_lowercase , i + 1 ) else: snake_case_ : Dict = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: snake_case_ : Dict = all_hidden_states + (hidden_states,) snake_case_ : str = (hidden_states,) if self.output_hidden_states: snake_case_ : List[Any] = outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : Union[str, Any] = outputs + (all_attentions,) snake_case_ : List[str] = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( '''The Bert Model transformer with early exiting (DeeBERT). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> Union[str, Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config snake_case_ : int = BertEmbeddings(_lowercase ) snake_case_ : Tuple = DeeBertEncoder(_lowercase ) snake_case_ : int = BertPooler(_lowercase ) self.init_weights() def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' self.encoder.init_highway_pooler(self.pooler ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return self.embeddings.word_embeddings def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = value def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(_lowercase ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' if input_ids is not None and inputs_embeds is not None: raise ValueError("""You cannot specify both input_ids and inputs_embeds at the same time""" ) elif input_ids is not None: snake_case_ : Dict = input_ids.size() elif inputs_embeds is not None: snake_case_ : Optional[int] = inputs_embeds.size()[:-1] else: raise ValueError("""You have to specify either input_ids or inputs_embeds""" ) snake_case_ : int = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: snake_case_ : Dict = torch.ones(_lowercase , device=_lowercase ) if encoder_attention_mask is None: snake_case_ : Tuple = torch.ones(_lowercase , device=_lowercase ) if token_type_ids is None: snake_case_ : Any = torch.zeros(_lowercase , dtype=torch.long , device=_lowercase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. snake_case_ : torch.Tensor = self.get_extended_attention_mask(_lowercase , _lowercase , _lowercase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: snake_case_ : List[str] = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: snake_case_ : Any = encoder_attention_mask[:, None, None, :] snake_case_ : List[str] = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility snake_case_ : List[str] = (1.0 - encoder_extended_attention_mask) * -1_0000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] snake_case_ : int = self.get_head_mask(_lowercase , self.config.num_hidden_layers ) snake_case_ : List[str] = self.embeddings( input_ids=_lowercase , position_ids=_lowercase , token_type_ids=_lowercase , inputs_embeds=_lowercase ) snake_case_ : List[str] = self.encoder( _lowercase , attention_mask=_lowercase , head_mask=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , ) snake_case_ : Optional[Any] = encoder_outputs[0] snake_case_ : Union[str, Any] = self.pooler(_lowercase ) snake_case_ : Optional[Any] = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = message snake_case_ : str = exit_layer # start from 1! class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__() snake_case_ : str = BertPooler(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Dict = nn.Linear(config.hidden_size , config.num_labels ) def UpperCAmelCase__ ( self , _lowercase ) -> Dict: '''simple docstring''' snake_case_ : List[str] = encoder_outputs[0] snake_case_ : List[Any] = self.pooler(_lowercase ) # "return" pooler_output # BertModel snake_case_ : Tuple = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification snake_case_ : Union[str, Any] = bmodel_output[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : List[str] = self.classifier(_lowercase ) return logits, pooled_output @add_start_docstrings( '''Bert Model (with early exiting - DeeBERT) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config.num_labels snake_case_ : Tuple = config.num_hidden_layers snake_case_ : Any = DeeBertModel(_lowercase ) snake_case_ : Optional[int] = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Tuple = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> int: '''simple docstring''' snake_case_ : int = self.num_layers try: snake_case_ : Any = self.bert( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits snake_case_ : str = outputs[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : List[Any] = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : Optional[int] = e.message snake_case_ : Dict = e.exit_layer snake_case_ : Optional[Any] = outputs[0] if not self.training: snake_case_ : int = entropy(_lowercase ) snake_case_ : int = [] snake_case_ : List[str] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : Dict = [] for highway_exit in outputs[-1]: snake_case_ : List[Any] = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : List[Any] = MSELoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : List[str] = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : str = (loss,) + outputs if not self.training: snake_case_ : Optional[int] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : str = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
58
0
from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging _UpperCamelCase = logging.get_logger(__name__) class _lowerCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" UpperCAmelCase_ : str =["pixel_values"] def __init__( self , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = PILImageResampling.BICUBIC , UpperCAmelCase = True , UpperCAmelCase = None , UpperCAmelCase = True , UpperCAmelCase = 1 / 255 , UpperCAmelCase = True , UpperCAmelCase = IMAGENET_DEFAULT_MEAN , UpperCAmelCase = IMAGENET_DEFAULT_STD , **UpperCAmelCase , ) -> None: '''simple docstring''' super().__init__(**_lowercase ) __snake_case : Tuple = size if size is not None else {"""shortest_edge""": 224} __snake_case : Tuple = get_size_dict(_lowercase , default_to_square=_lowercase ) __snake_case : Tuple = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} __snake_case : Optional[int] = get_size_dict(_lowercase , param_name="crop_size" ) __snake_case : Union[str, Any] = do_resize __snake_case : List[Any] = size __snake_case : Optional[Any] = resample __snake_case : Tuple = do_center_crop __snake_case : Optional[Any] = crop_size __snake_case : List[str] = do_rescale __snake_case : int = rescale_factor __snake_case : Tuple = do_normalize __snake_case : str = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __snake_case : Any = image_std if image_std is not None else IMAGENET_DEFAULT_STD def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = PILImageResampling.BICUBIC , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: '''simple docstring''' __snake_case : List[Any] = get_size_dict(_lowercase , default_to_square=_lowercase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __snake_case : Optional[Any] = int((256 / 224) * size["shortest_edge"] ) __snake_case : Union[str, Any] = get_resize_output_image_size(_lowercase , size=_lowercase , default_to_square=_lowercase ) __snake_case : Tuple = {"""height""": output_size[0], """width""": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"""Size dict must have keys \'height\' and \'width\' or \'shortest_edge\'. Got {size_dict.keys()}""" ) return resize( _lowercase , size=(size_dict["height"], size_dict["width"]) , resample=_lowercase , data_format=_lowercase , **_lowercase ) def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: '''simple docstring''' __snake_case : str = get_size_dict(_lowercase ) if "height" not in size or "width" not in size: raise ValueError(F"""Size dict must have keys \'height\' and \'width\'. Got {size.keys()}""" ) return center_crop(_lowercase , size=(size["height"], size["width"]) , data_format=_lowercase , **_lowercase ) def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: '''simple docstring''' return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = None , **UpperCAmelCase , ) -> np.ndarray: '''simple docstring''' return normalize(_lowercase , mean=_lowercase , std=_lowercase , data_format=_lowercase , **_lowercase ) def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = None , UpperCAmelCase = ChannelDimension.FIRST , **UpperCAmelCase , ) -> BatchFeature: '''simple docstring''' __snake_case : List[str] = do_resize if do_resize is not None else self.do_resize __snake_case : str = resample if resample is not None else self.resample __snake_case : Dict = do_center_crop if do_center_crop is not None else self.do_center_crop __snake_case : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Any = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : List[str] = do_normalize if do_normalize is not None else self.do_normalize __snake_case : Optional[int] = image_mean if image_mean is not None else self.image_mean __snake_case : Any = image_std if image_std is not None else self.image_std __snake_case : Optional[int] = size if size is not None else self.size __snake_case : List[str] = get_size_dict(_lowercase , default_to_square=_lowercase ) __snake_case : Any = crop_size if crop_size is not None else self.crop_size __snake_case : Any = get_size_dict(_lowercase , param_name="crop_size" ) __snake_case : Optional[Any] = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __snake_case : List[str] = [to_numpy_array(_lowercase ) for image in images] if do_resize: __snake_case : Any = [self.resize(_lowercase , _lowercase , _lowercase ) for image in images] if do_center_crop: __snake_case : Dict = [self.center_crop(_lowercase , _lowercase ) for image in images] if do_rescale: __snake_case : str = [self.rescale(_lowercase , _lowercase ) for image in images] if do_normalize: __snake_case : List[str] = [self.normalize(_lowercase , _lowercase , _lowercase ) for image in images] __snake_case : List[str] = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] __snake_case : List[Any] = {"""pixel_values""": images} return BatchFeature(data=_lowercase , tensor_type=_lowercase )
243
"""simple docstring""" # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' return 1 / (1 + np.exp(-z )) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int ): '''simple docstring''' return (-y * np.log(__UpperCamelCase ) - (1 - y) * np.log(1 - h )).mean() def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : List[Any] , __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Optional[int] = np.dot(__UpperCamelCase , __UpperCamelCase ) return np.sum(y * scores - np.log(1 + np.exp(__UpperCamelCase ) ) ) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : int=7_0_0_0_0 ): '''simple docstring''' snake_case_ : Dict = np.zeros(x.shape[1] ) for iterations in range(__UpperCamelCase ): snake_case_ : Any = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Optional[Any] = np.dot(x.T , h - y ) / y.size snake_case_ : str = theta - alpha * gradient # updating the weights snake_case_ : int = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Dict = cost_function(__UpperCamelCase , __UpperCamelCase ) if iterations % 1_0_0 == 0: print(F'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": __lowerCAmelCase : Any = datasets.load_iris() __lowerCAmelCase : List[Any] = iris.data[:, :2] __lowerCAmelCase : Tuple = (iris.target != 0) * 1 __lowerCAmelCase : Any = 0.1 __lowerCAmelCase : List[Any] = logistic_reg(alpha, x, y, max_iterations=7_0000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' return sigmoid_function( np.dot(__UpperCamelCase , __UpperCamelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = (x[:, 0].min(), x[:, 0].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Tuple = (x[:, 1].min(), x[:, 1].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Optional[Any] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) __lowerCAmelCase : Any = np.c_[xxa.ravel(), xxa.ravel()] __lowerCAmelCase : Optional[int] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
58
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available UpperCamelCase_ : Optional[int] = { '''configuration_ernie''': ['''ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ErnieConfig''', '''ErnieOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ : Optional[Any] = [ '''ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ErnieForCausalLM''', '''ErnieForMaskedLM''', '''ErnieForMultipleChoice''', '''ErnieForNextSentencePrediction''', '''ErnieForPreTraining''', '''ErnieForQuestionAnswering''', '''ErnieForSequenceClassification''', '''ErnieForTokenClassification''', '''ErnieModel''', '''ErniePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys UpperCamelCase_ : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
115
"""simple docstring""" import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput __lowerCAmelCase : Tuple = '''scheduler_config.json''' class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 1 _lowerCamelCase = 2 _lowerCamelCase = 3 _lowerCamelCase = 4 _lowerCamelCase = 5 @dataclass class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 42 class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = SCHEDULER_CONFIG_NAME _lowerCamelCase = ['''dtype'''] _lowerCamelCase = [] _lowerCamelCase = True @classmethod def UpperCAmelCase__ ( cls , _lowercase = None , _lowercase = None , _lowercase=False , **_lowercase , ) -> Any: '''simple docstring''' snake_case_ , snake_case_ : int = cls.load_config( pretrained_model_name_or_path=_lowercase , subfolder=_lowercase , return_unused_kwargs=_lowercase , **_lowercase , ) snake_case_ , snake_case_ : Dict = cls.from_config(_lowercase , return_unused_kwargs=_lowercase , **_lowercase ) if hasattr(_lowercase , """create_state""" ) and getattr(_lowercase , """has_state""" , _lowercase ): snake_case_ : Any = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCAmelCase__ ( self , _lowercase , _lowercase = False , **_lowercase ) -> Optional[Any]: '''simple docstring''' self.save_config(save_directory=_lowercase , push_to_hub=_lowercase , **_lowercase ) @property def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' return self._get_compatibles() @classmethod def UpperCAmelCase__ ( cls ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = list(set([cls.__name__] + cls._compatibles ) ) snake_case_ : str = importlib.import_module(__name__.split(""".""" )[0] ) snake_case_ : Optional[int] = [ getattr(_lowercase , _lowercase ) for c in compatible_classes_str if hasattr(_lowercase , _lowercase ) ] return compatible_classes def __lowerCAmelCase ( __UpperCamelCase : jnp.ndarray , __UpperCamelCase : Tuple[int] ): '''simple docstring''' assert len(__UpperCamelCase ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__UpperCamelCase ) - x.ndim) ) , __UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Any=0.999 , __UpperCamelCase : Optional[int]=jnp.floataa ): '''simple docstring''' def alpha_bar(__UpperCamelCase : Optional[int] ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 snake_case_ : Optional[Any] = [] for i in range(__UpperCamelCase ): snake_case_ : Dict = i / num_diffusion_timesteps snake_case_ : Union[str, Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__UpperCamelCase ) / alpha_bar(__UpperCamelCase ) , __UpperCamelCase ) ) return jnp.array(__UpperCamelCase , dtype=__UpperCamelCase ) @flax.struct.dataclass class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 @classmethod def UpperCAmelCase__ ( cls , _lowercase ) -> int: '''simple docstring''' snake_case_ : Any = scheduler.config if config.trained_betas is not None: snake_case_ : Optional[Any] = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": snake_case_ : int = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case_ : str = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case_ : int = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( f'beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}' ) snake_case_ : Optional[Any] = 1.0 - betas snake_case_ : Any = jnp.cumprod(_lowercase , axis=0 ) return cls( alphas=_lowercase , betas=_lowercase , alphas_cumprod=_lowercase , ) def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ : Tuple = state.alphas_cumprod snake_case_ : Optional[int] = alphas_cumprod[timesteps] ** 0.5 snake_case_ : Dict = sqrt_alpha_prod.flatten() snake_case_ : int = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) snake_case_ : Optional[Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 snake_case_ : Dict = sqrt_one_minus_alpha_prod.flatten() snake_case_ : Tuple = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : str = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : List[Any] = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
58
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = {'''configuration_mmbt''': ['''MMBTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ['''MMBTForClassification''', '''MMBTModel''', '''ModalEmbeddings'''] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
74
"""simple docstring""" from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( '''The RoBERTa Model transformer with early exiting (DeeRoBERTa). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : str = RobertaEmbeddings(_lowercase ) self.init_weights() @add_start_docstrings( '''RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Optional[Any] = config.num_labels snake_case_ : Dict = config.num_hidden_layers snake_case_ : str = DeeRobertaModel(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : List[str] = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> Tuple: '''simple docstring''' snake_case_ : Any = self.num_layers try: snake_case_ : int = self.roberta( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) snake_case_ : str = outputs[1] snake_case_ : Union[str, Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : Dict = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : List[Any] = e.message snake_case_ : Union[str, Any] = e.exit_layer snake_case_ : Dict = outputs[0] if not self.training: snake_case_ : Dict = entropy(_lowercase ) snake_case_ : Optional[int] = [] snake_case_ : Union[str, Any] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Dict = MSELoss() snake_case_ : Dict = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Union[str, Any] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : int = [] for highway_exit in outputs[-1]: snake_case_ : Tuple = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Optional[Any] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Optional[int] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : Dict = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : List[str] = (loss,) + outputs if not self.training: snake_case_ : Optional[Any] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : Tuple = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
58
0
'''simple docstring''' import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): lowercase : Union[str, Any] =0 lowercase : Optional[Any] =False lowercase : Optional[Any] =3.0 class __UpperCamelCase ( unittest.TestCase ): def lowercase__ ( self ): """simple docstring""" self.assertDictEqual(MockClass().to_kwargs(), {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs(), {'''a''': 2} ) self.assertDictEqual(MockClass(a=2, b=_lowercase ).to_kwargs(), {'''a''': 2, '''b''': True} ) self.assertDictEqual(MockClass(a=2, c=2.2_5 ).to_kwargs(), {'''a''': 2, '''c''': 2.2_5} ) @require_cuda def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =GradScalerKwargs(init_scale=1_024, growth_factor=2 ) AcceleratorState._reset_state() lowerCamelCase_ =Accelerator(mixed_precision='''fp16''', kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) lowerCamelCase_ =accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale, 1_0_2_4.0 ) self.assertEqual(scaler._growth_factor, 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor, 0.5 ) self.assertEqual(scaler._growth_interval, 2_000 ) self.assertEqual(scaler._enabled, _lowercase ) @require_multi_gpu def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] execute_subprocess_async(_lowercase, env=os.environ.copy() ) if __name__ == "__main__": a_ : Any = DistributedDataParallelKwargs(bucket_cap_mb=15, find_unused_parameters=True) a_ : Tuple = Accelerator(kwargs_handlers=[ddp_scaler]) a_ : Optional[Any] = torch.nn.Linear(1_00, 2_00) a_ : Union[str, Any] = accelerator.prepare(model) # Check the values changed in kwargs a_ : Optional[Any] = '''''' a_ : str = model.bucket_bytes_cap // (10_24 * 10_24) if observed_bucket_cap_map != 15: error_msg += F"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += F"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += F"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += F"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += F"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
676
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : list[int] ): '''simple docstring''' if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : list[int] , __UpperCamelCase : int ): '''simple docstring''' if curr_ind == len(__UpperCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(__UpperCamelCase ) ): if valid_connection(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): # Insert current vertex into path as next transition snake_case_ : List[str] = next_ver # Validate created path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , curr_ind + 1 ): return True # Backtrack snake_case_ : Tuple = -1 return False def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int = 0 ): '''simple docstring''' snake_case_ : Tuple = [-1] * (len(__UpperCamelCase ) + 1) # initialize start and end of path with starting index snake_case_ : Optional[int] = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , 1 ) else []
58
0
import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def _A ( SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE ): UpperCAmelCase__: List[Any] = FunnelConfig.from_json_file(__UpperCamelCase ) print(f"Building PyTorch model from configuration: {config}" ) UpperCAmelCase__: Dict = FunnelBaseModel(__UpperCamelCase ) if base_model else FunnelModel(__UpperCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # Save pytorch-model print(f"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() ,__UpperCamelCase ) if __name__ == "__main__": _lowerCAmelCase : str =argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--base_model""", action="""store_true""", help="""Whether you want just the base model (no decoder) or not.""" ) _lowerCAmelCase : Dict =parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
113
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = ['''image_processor''', '''tokenizer'''] _lowerCamelCase = '''BlipImageProcessor''' _lowerCamelCase = '''AutoTokenizer''' def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__(_lowercase , _lowercase ) # add QFormer tokenizer snake_case_ : List[str] = qformer_tokenizer def __call__( self , _lowercase = None , _lowercase = None , _lowercase = True , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = 0 , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = True , _lowercase = None , **_lowercase , ) -> BatchFeature: '''simple docstring''' if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) snake_case_ : Optional[Any] = BatchFeature() if text is not None: snake_case_ : List[str] = self.tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) encoding.update(_lowercase ) snake_case_ : Union[str, Any] = self.qformer_tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) snake_case_ : List[str] = qformer_text_encoding.pop("""input_ids""" ) snake_case_ : Union[str, Any] = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: snake_case_ : Tuple = self.image_processor(_lowercase , return_tensors=_lowercase ) encoding.update(_lowercase ) return encoding def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> List[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> Dict: '''simple docstring''' return self.tokenizer.decode(*_lowercase , **_lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' snake_case_ : str = self.tokenizer.model_input_names snake_case_ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def UpperCAmelCase__ ( self , _lowercase , **_lowercase ) -> Optional[int]: '''simple docstring''' if os.path.isfile(_lowercase ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(_lowercase , exist_ok=_lowercase ) snake_case_ : int = os.path.join(_lowercase , """qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(_lowercase ) return super().save_pretrained(_lowercase , **_lowercase ) @classmethod def UpperCAmelCase__ ( cls , _lowercase , **_lowercase ) -> int: '''simple docstring''' snake_case_ : List[str] = AutoTokenizer.from_pretrained(_lowercase , subfolder="""qformer_tokenizer""" ) snake_case_ : Union[str, Any] = cls._get_arguments_from_pretrained(_lowercase , **_lowercase ) args.append(_lowercase ) return cls(*_lowercase )
58
0
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging __UpperCamelCase : Optional[int] = logging.get_logger(__name__) __UpperCamelCase : Tuple = { '''EleutherAI/gpt-j-6B''': '''https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json''', # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowercase__ = "gptj" lowercase__ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : int ,lowercase_ : List[str]=5_0_4_0_0 ,lowercase_ : Union[str, Any]=2_0_4_8 ,lowercase_ : Union[str, Any]=4_0_9_6 ,lowercase_ : List[Any]=2_8 ,lowercase_ : Dict=1_6 ,lowercase_ : List[str]=6_4 ,lowercase_ : List[str]=None ,lowercase_ : List[str]="gelu_new" ,lowercase_ : List[Any]=0.0 ,lowercase_ : List[Any]=0.0 ,lowercase_ : Union[str, Any]=0.0 ,lowercase_ : List[Any]=1E-5 ,lowercase_ : str=0.02 ,lowercase_ : Tuple=True ,lowercase_ : Any=5_0_2_5_6 ,lowercase_ : int=5_0_2_5_6 ,lowercase_ : Union[str, Any]=False ,**lowercase_ : Union[str, Any] ,): lowerCAmelCase__ : Dict = vocab_size lowerCAmelCase__ : str = n_positions lowerCAmelCase__ : List[str] = n_embd lowerCAmelCase__ : List[str] = n_layer lowerCAmelCase__ : str = n_head lowerCAmelCase__ : List[Any] = n_inner lowerCAmelCase__ : Union[str, Any] = rotary_dim lowerCAmelCase__ : str = activation_function lowerCAmelCase__ : Tuple = resid_pdrop lowerCAmelCase__ : str = embd_pdrop lowerCAmelCase__ : Optional[int] = attn_pdrop lowerCAmelCase__ : Tuple = layer_norm_epsilon lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : Dict = use_cache lowerCAmelCase__ : Any = bos_token_id lowerCAmelCase__ : Dict = eos_token_id super().__init__( bos_token_id=_lowercase ,eos_token_id=_lowercase ,tie_word_embeddings=_lowercase ,**_lowercase ) class SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self : str ,lowercase_ : List[str] ,lowercase_ : Union[str, Any] = "default" ,lowercase_ : Union[str, Any] = None ,lowercase_ : Union[str, Any] = False ,): super().__init__(_lowercase ,task=_lowercase ,patching_specs=_lowercase ,use_past=_lowercase ) if not getattr(self._config ,'''pad_token_id''' ,_lowercase ): # TODO: how to do that better? lowerCAmelCase__ : Dict = 0 @property def __lowerCAmelCase ( self : Optional[Any] ): lowerCAmelCase__ : Dict = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: self.fill_with_past_key_values_(_lowercase ,direction='''inputs''' ) lowerCAmelCase__ : int = {0: """batch""", 1: """past_sequence + sequence"""} else: lowerCAmelCase__ : Optional[int] = {0: """batch""", 1: """sequence"""} return common_inputs @property def __lowerCAmelCase ( self : str ): return self._config.n_layer @property def __lowerCAmelCase ( self : Optional[Any] ): return self._config.n_head def __lowerCAmelCase ( self : Tuple ,lowercase_ : Any ,lowercase_ : Optional[int] = -1 ,lowercase_ : Union[str, Any] = -1 ,lowercase_ : Optional[Any] = False ,lowercase_ : Dict = None ,): lowerCAmelCase__ : List[str] = super(_lowercase ,self ).generate_dummy_inputs( _lowercase ,batch_size=_lowercase ,seq_length=_lowercase ,is_pair=_lowercase ,framework=_lowercase ) # We need to order the input in the way they appears in the forward() lowerCAmelCase__ : Optional[Any] = OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowerCAmelCase__ : Dict = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values lowerCAmelCase__ : List[Any] = seqlen + 2 lowerCAmelCase__ : Dict = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCAmelCase__ : str = [ (torch.zeros(_lowercase ), torch.zeros(_lowercase )) for _ in range(self.num_layers ) ] lowerCAmelCase__ : Dict = common_inputs["""attention_mask"""] if self.use_past: lowerCAmelCase__ : List[Any] = ordered_inputs["""attention_mask"""].dtype lowerCAmelCase__ : Union[str, Any] = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(_lowercase ,_lowercase ,dtype=_lowercase )] ,dim=1 ) return ordered_inputs @property def __lowerCAmelCase ( self : Optional[Any] ): return 1_3
450
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCAmelCase : List[Any] = { '''configuration_vivit''': ['''VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VivitConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = ['''VivitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = [ '''VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VivitModel''', '''VivitPreTrainedModel''', '''VivitForVideoClassification''', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
58
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable lowerCamelCase_ = {'''configuration_dpt''': ['''DPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''DPTConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = ['''DPTFeatureExtractor'''] lowerCamelCase_ = ['''DPTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''DPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''DPTForDepthEstimation''', '''DPTForSemanticSegmentation''', '''DPTModel''', '''DPTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_dpt import DPT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPTConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_dpt import DPTFeatureExtractor from .image_processing_dpt import DPTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dpt import ( DPT_PRETRAINED_MODEL_ARCHIVE_LIST, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel, DPTPreTrainedModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
318
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() __lowerCAmelCase : Dict = logging.get_logger(__name__) def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Dict , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : List[str] = WavaVecaForSequenceClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : int = downstream_dict["""projector.weight"""] snake_case_ : Optional[int] = downstream_dict["""projector.bias"""] snake_case_ : List[Any] = downstream_dict["""model.post_net.linear.weight"""] snake_case_ : Union[str, Any] = downstream_dict["""model.post_net.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Tuple , __UpperCamelCase : str ): '''simple docstring''' snake_case_ : int = WavaVecaForAudioFrameClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""model.linear.weight"""] snake_case_ : int = downstream_dict["""model.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[str] , __UpperCamelCase : List[Any] ): '''simple docstring''' snake_case_ : Optional[int] = WavaVecaForXVector.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""connector.weight"""] snake_case_ : str = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): snake_case_ : Dict = downstream_dict[ F'model.framelevel_feature_extractor.module.{i}.kernel.weight' ] snake_case_ : int = downstream_dict[F'model.framelevel_feature_extractor.module.{i}.kernel.bias'] snake_case_ : str = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] snake_case_ : Any = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] snake_case_ : List[str] = downstream_dict["""objective.W"""] return model @torch.no_grad() def __lowerCAmelCase ( __UpperCamelCase : List[Any] , __UpperCamelCase : Any , __UpperCamelCase : List[str] , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : Any = torch.load(__UpperCamelCase , map_location="""cpu""" ) snake_case_ : Any = checkpoint["""Downstream"""] snake_case_ : Optional[Any] = WavaVecaConfig.from_pretrained(__UpperCamelCase ) snake_case_ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained( __UpperCamelCase , return_attention_mask=__UpperCamelCase , do_normalize=__UpperCamelCase ) snake_case_ : Optional[Any] = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): snake_case_ : Tuple = convert_classification(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForAudioFrameClassification""" ): snake_case_ : Union[str, Any] = convert_diarization(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForXVector""" ): snake_case_ : List[str] = convert_xvector(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: raise NotImplementedError(F'S3PRL weights conversion is not supported for {arch}' ) if hf_config.use_weighted_layer_sum: snake_case_ : List[Any] = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(__UpperCamelCase ) hf_model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __lowerCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--base_model_name''', default=None, type=str, help='''Name of the huggingface pretrained base model.''' ) parser.add_argument('''--config_path''', default=None, type=str, help='''Path to the huggingface classifier config.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to the s3prl checkpoint.''') parser.add_argument('''--model_dump_path''', default=None, type=str, help='''Path to the final converted model.''') __lowerCAmelCase : Dict = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
58
0
"""simple docstring""" class _UpperCAmelCase: def __init__( self , __a , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = None _UpperCamelCase = None _UpperCamelCase = graph self._normalize_graph(_lowercase , _lowercase) _UpperCamelCase = len(_lowercase) _UpperCamelCase = None def UpperCAmelCase ( self , __a , __a) -> Tuple: '''simple docstring''' if sources is int: _UpperCamelCase = [sources] if sinks is int: _UpperCamelCase = [sinks] if len(_lowercase) == 0 or len(_lowercase) == 0: return _UpperCamelCase = sources[0] _UpperCamelCase = sinks[0] # make fake vertex if there are more # than one source or sink if len(_lowercase) > 1 or len(_lowercase) > 1: _UpperCamelCase = 0 for i in sources: max_input_flow += sum(self.graph[i]) _UpperCamelCase = len(self.graph) + 1 for room in self.graph: room.insert(0 , 0) self.graph.insert(0 , [0] * size) for i in sources: _UpperCamelCase = max_input_flow _UpperCamelCase = 0 _UpperCamelCase = len(self.graph) + 1 for room in self.graph: room.append(0) self.graph.append([0] * size) for i in sinks: _UpperCamelCase = max_input_flow _UpperCamelCase = size - 1 def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' if self.maximum_flow_algorithm is None: raise Exception('''You need to set maximum flow algorithm before.''') if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = algorithm(self) class _UpperCAmelCase: def __init__( self , __a) -> Any: '''simple docstring''' _UpperCamelCase = flow_network _UpperCamelCase = flow_network.verticesCount _UpperCamelCase = flow_network.sourceIndex _UpperCamelCase = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that _UpperCamelCase = flow_network.graph _UpperCamelCase = False def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' if not self.executed: self._algorithm() _UpperCamelCase = True def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' pass class _UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): def __init__( self , __a) -> List[Any]: '''simple docstring''' super().__init__(_lowercase) # use this to save your result _UpperCamelCase = -1 def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' if not self.executed: raise Exception('''You should execute algorithm before using its result!''') return self.maximum_flow class _UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): def __init__( self , __a) -> Optional[int]: '''simple docstring''' super().__init__(_lowercase) _UpperCamelCase = [[0] * self.verticies_count for i in range(self.verticies_count)] _UpperCamelCase = [0] * self.verticies_count _UpperCamelCase = [0] * self.verticies_count def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index]): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule _UpperCamelCase = [ i for i in range(self.verticies_count) if i != self.source_index and i != self.sink_index ] # move through list _UpperCamelCase = 0 while i < len(_lowercase): _UpperCamelCase = vertices_list[i] _UpperCamelCase = self.heights[vertex_index] self.process_vertex(_lowercase) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(_lowercase)) _UpperCamelCase = 0 else: i += 1 _UpperCamelCase = sum(self.preflow[self.source_index]) def UpperCAmelCase ( self , __a) -> Optional[int]: '''simple docstring''' while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(_lowercase , _lowercase) self.relabel(_lowercase) def UpperCAmelCase ( self , __a , __a) -> str: '''simple docstring''' _UpperCamelCase = min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def UpperCAmelCase ( self , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = None for to_index in range(self.verticies_count): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): _UpperCamelCase = self.heights[to_index] if min_height is not None: _UpperCamelCase = min_height + 1 if __name__ == "__main__": _a = [0] _a = [3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] _a = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network _a = FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate _a = flow_network.find_maximum_flow() print(F"""maximum flow is {maximum_flow}""")
19
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __lowerCAmelCase : Tuple = logging.get_logger(__name__) __lowerCAmelCase : int = {'''vocab_file''': '''vocab.txt'''} __lowerCAmelCase : Union[str, Any] = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } __lowerCAmelCase : Optional[Any] = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } __lowerCAmelCase : Any = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = PRETRAINED_INIT_CONFIGURATION _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase = ConvBertTokenizer def __init__( self , _lowercase=None , _lowercase=None , _lowercase=True , _lowercase="[UNK]" , _lowercase="[SEP]" , _lowercase="[PAD]" , _lowercase="[CLS]" , _lowercase="[MASK]" , _lowercase=True , _lowercase=None , **_lowercase , ) -> Union[str, Any]: '''simple docstring''' super().__init__( _lowercase , tokenizer_file=_lowercase , do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , tokenize_chinese_chars=_lowercase , strip_accents=_lowercase , **_lowercase , ) snake_case_ : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowercase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowercase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowercase ) != tokenize_chinese_chars ): snake_case_ : Optional[int] = getattr(_lowercase , normalizer_state.pop("""type""" ) ) snake_case_ : Dict = do_lower_case snake_case_ : str = strip_accents snake_case_ : Optional[Any] = tokenize_chinese_chars snake_case_ : int = normalizer_class(**_lowercase ) snake_case_ : Optional[int] = do_lower_case def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> int: '''simple docstring''' snake_case_ : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> List[int]: '''simple docstring''' snake_case_ : int = [self.sep_token_id] snake_case_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[int] = self._tokenizer.model.save(_lowercase , name=_lowercase ) return tuple(_lowercase )
58
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : List[str] = { '''configuration_time_series_transformer''': [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimeSeriesTransformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Tuple = [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimeSeriesTransformerForPrediction''', '''TimeSeriesTransformerModel''', '''TimeSeriesTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys UpperCAmelCase : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
627
"""simple docstring""" import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" @register_to_config def __init__( self , _lowercase = 1_2_8 , _lowercase = 2_5_6 , _lowercase = 2000.0 , _lowercase = 7_6_8 , _lowercase = 1_2 , _lowercase = 1_2 , _lowercase = 6_4 , _lowercase = 2_0_4_8 , _lowercase = 0.1 , ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Optional[Any] = nn.Sequential( nn.Linear(_lowercase , d_model * 4 , bias=_lowercase ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_lowercase ) , nn.SiLU() , ) snake_case_ : Any = nn.Embedding(_lowercase , _lowercase ) snake_case_ : Union[str, Any] = False snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Union[str, Any] = nn.Dropout(p=_lowercase ) snake_case_ : Tuple = nn.ModuleList() for lyr_num in range(_lowercase ): # FiLM conditional T5 decoder snake_case_ : Union[str, Any] = DecoderLayer(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) self.decoders.append(_lowercase ) snake_case_ : List[Any] = TaLayerNorm(_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(p=_lowercase ) snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def UpperCAmelCase__ ( self , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ , snake_case_ , snake_case_ : str = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. snake_case_ : Optional[int] = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) snake_case_ : int = self.conditioning_emb(_lowercase ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) snake_case_ : Tuple = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. snake_case_ : Dict = torch.broadcast_to( torch.arange(_lowercase , device=decoder_input_tokens.device ) , (batch, seq_length) , ) snake_case_ : Tuple = self.position_encoding(_lowercase ) snake_case_ : Optional[Any] = self.continuous_inputs_projection(_lowercase ) inputs += position_encodings snake_case_ : List[Any] = self.dropout(_lowercase ) # decoder: No padding present. snake_case_ : Tuple = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. snake_case_ : int = [(x, self.encoder_decoder_mask(_lowercase , _lowercase )) for x, y in encodings_and_masks] # cross attend style: concat encodings snake_case_ : Optional[Any] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) snake_case_ : str = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: snake_case_ : int = lyr( _lowercase , conditioning_emb=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , )[0] snake_case_ : int = self.decoder_norm(_lowercase ) snake_case_ : Union[str, Any] = self.post_dropout(_lowercase ) snake_case_ : int = self.spec_out(_lowercase ) return spec_out class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase=1E-6 ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Any = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase ) ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = self.layer[0]( _lowercase , conditioning_emb=_lowercase , attention_mask=_lowercase , ) if encoder_hidden_states is not None: snake_case_ : Tuple = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) snake_case_ : str = self.layer[1]( _lowercase , key_value_states=_lowercase , attention_mask=_lowercase , ) # Apply Film Conditional Feed Forward layer snake_case_ : Any = self.layer[-1](_lowercase , _lowercase ) return (hidden_states,) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> str: '''simple docstring''' super().__init__() snake_case_ : Any = TaLayerNorm(_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Union[str, Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : List[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : str = self.FiLMLayer(_lowercase , _lowercase ) # Self-attention block snake_case_ : List[Any] = self.attention(_lowercase ) snake_case_ : List[str] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__() snake_case_ : List[Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : Union[str, Any] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) snake_case_ : Optional[Any] = self.attention( _lowercase , encoder_hidden_states=_lowercase , attention_mask=attention_mask.squeeze(1 ) , ) snake_case_ : Any = hidden_states + self.dropout(_lowercase ) return layer_output class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Tuple = TaDenseGatedActDense(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Optional[int] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Tuple = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> str: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : Optional[int] = self.film(_lowercase , _lowercase ) snake_case_ : int = self.DenseReluDense(_lowercase ) snake_case_ : Optional[Any] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__() snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Any = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : int = nn.Dropout(_lowercase ) snake_case_ : Optional[int] = NewGELUActivation() def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' snake_case_ : str = self.act(self.wi_a(_lowercase ) ) snake_case_ : Dict = self.wi_a(_lowercase ) snake_case_ : Any = hidden_gelu * hidden_linear snake_case_ : List[Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.wo(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1E-6 ) -> str: '''simple docstring''' super().__init__() snake_case_ : Union[str, Any] = nn.Parameter(torch.ones(_lowercase ) ) snake_case_ : int = eps def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=_lowercase ) snake_case_ : Any = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: snake_case_ : str = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def UpperCAmelCase__ ( self , _lowercase ) -> torch.Tensor: '''simple docstring''' return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.04_4715 * torch.pow(_lowercase , 3.0 )) )) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Any: '''simple docstring''' super().__init__() snake_case_ : List[Any] = nn.Linear(_lowercase , out_features * 2 , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = self.scale_bias(_lowercase ) snake_case_ , snake_case_ : Any = torch.chunk(_lowercase , 2 , -1 ) snake_case_ : Optional[Any] = x * (1 + scale) + shift return x
58
0
def lowerCamelCase__ ( snake_case_ : list[list[int]] , snake_case_ : int , snake_case_ : int , snake_case_ : list[int] ) -> Dict: if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def lowerCamelCase__ ( snake_case_ : list[list[int]] , snake_case_ : list[int] , snake_case_ : int ) -> Tuple: if curr_ind == len(__UpperCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(__UpperCamelCase ) ): if valid_connection(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): # Insert current vertex into path as next transition __snake_case = next_ver # Validate created path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , curr_ind + 1 ): return True # Backtrack __snake_case = -1 return False def lowerCamelCase__ ( snake_case_ : list[list[int]] , snake_case_ : int = 0 ) -> Tuple: __snake_case = [-1] * (len(__UpperCamelCase ) + 1) # initialize start and end of path with starting index __snake_case = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , 1 ) else []
592
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''roformer''' def __init__( self , _lowercase=5_0_0_0_0 , _lowercase=None , _lowercase=7_6_8 , _lowercase=1_2 , _lowercase=1_2 , _lowercase=3_0_7_2 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=1_5_3_6 , _lowercase=2 , _lowercase=0.02 , _lowercase=1E-12 , _lowercase=0 , _lowercase=False , _lowercase=True , **_lowercase , ) -> int: '''simple docstring''' super().__init__(pad_token_id=_lowercase , **_lowercase ) snake_case_ : str = vocab_size snake_case_ : Any = hidden_size if embedding_size is None else embedding_size snake_case_ : List[str] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : Tuple = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : List[str] = rotary_value snake_case_ : str = use_cache class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ : Any = {0: """batch""", 1: """sequence"""} snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
58
0
"""simple docstring""" import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { '''google/owlvit-base-patch32''': '''https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json''', '''google/owlvit-base-patch16''': '''https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json''', '''google/owlvit-large-patch14''': '''https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): _lowerCAmelCase = "owlvit_text_model" def __init__(self , _lowercase=49408 , _lowercase=512 , _lowercase=2048 , _lowercase=12 , _lowercase=8 , _lowercase=16 , _lowercase="quick_gelu" , _lowercase=1e-5 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1.0 , _lowercase=0 , _lowercase=49406 , _lowercase=49407 , **_lowercase , ): '''simple docstring''' super().__init__(pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) __a : int = vocab_size __a : Tuple = hidden_size __a : Any = intermediate_size __a : List[Any] = num_hidden_layers __a : str = num_attention_heads __a : Dict = max_position_embeddings __a : Union[str, Any] = hidden_act __a : Any = layer_norm_eps __a : Union[str, Any] = attention_dropout __a : Any = initializer_range __a : int = initializer_factor @classmethod def lowerCAmelCase__(cls , _lowercase , **_lowercase ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) __a : Union[str, Any] = cls.get_config_dict(_lowercase , **_lowercase ) # get the text config dict if we are loading from OwlViTConfig if config_dict.get("""model_type""" ) == "owlvit": __a : str = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): _lowerCAmelCase = "owlvit_vision_model" def __init__(self , _lowercase=768 , _lowercase=3072 , _lowercase=12 , _lowercase=12 , _lowercase=3 , _lowercase=768 , _lowercase=32 , _lowercase="quick_gelu" , _lowercase=1e-5 , _lowercase=0.0 , _lowercase=0.02 , _lowercase=1.0 , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) __a : Optional[int] = hidden_size __a : Optional[int] = intermediate_size __a : Union[str, Any] = num_hidden_layers __a : str = num_attention_heads __a : Union[str, Any] = num_channels __a : int = image_size __a : Dict = patch_size __a : Tuple = hidden_act __a : Tuple = layer_norm_eps __a : Optional[int] = attention_dropout __a : str = initializer_range __a : Union[str, Any] = initializer_factor @classmethod def lowerCAmelCase__(cls , _lowercase , **_lowercase ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) __a : int = cls.get_config_dict(_lowercase , **_lowercase ) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get("""model_type""" ) == "owlvit": __a : Any = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): _lowerCAmelCase = "owlvit" _lowerCAmelCase = True def __init__(self , _lowercase=None , _lowercase=None , _lowercase=512 , _lowercase=2.6592 , _lowercase=True , **_lowercase , ): '''simple docstring''' super().__init__(**_lowercase ) if text_config is None: __a : Optional[Any] = {} logger.info("""text_config is None. Initializing the OwlViTTextConfig with default values.""" ) if vision_config is None: __a : Tuple = {} logger.info("""vision_config is None. initializing the OwlViTVisionConfig with default values.""" ) __a : Optional[int] = OwlViTTextConfig(**_lowercase ) __a : Optional[int] = OwlViTVisionConfig(**_lowercase ) __a : Union[str, Any] = projection_dim __a : str = logit_scale_init_value __a : int = return_dict __a : List[str] = 1.0 @classmethod def lowerCAmelCase__(cls , _lowercase , **_lowercase ): '''simple docstring''' cls._set_token_in_kwargs(_lowercase ) __a : List[str] = cls.get_config_dict(_lowercase , **_lowercase ) if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_lowercase , **_lowercase ) @classmethod def lowerCAmelCase__(cls , _lowercase , _lowercase , **_lowercase ): '''simple docstring''' __a : List[str] = {} __a : List[str] = text_config __a : Any = vision_config return cls.from_dict(_lowercase , **_lowercase ) def lowerCAmelCase__(self ): '''simple docstring''' __a : List[Any] = copy.deepcopy(self.__dict__ ) __a : Dict = self.text_config.to_dict() __a : Optional[int] = self.vision_config.to_dict() __a : int = self.__class__.model_type return output class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): @property def lowerCAmelCase__(self ): '''simple docstring''' return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""attention_mask""", {0: """batch""", 1: """sequence"""}), ] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return OrderedDict( [ ("""logits_per_image""", {0: """batch"""}), ("""logits_per_text""", {0: """batch"""}), ("""text_embeds""", {0: """batch"""}), ("""image_embeds""", {0: """batch"""}), ] ) @property def lowerCAmelCase__(self ): '''simple docstring''' return 1e-4 def lowerCAmelCase__(self , _lowercase , _lowercase = -1 , _lowercase = -1 , _lowercase = None , ): '''simple docstring''' __a : Any = super().generate_dummy_inputs( processor.tokenizer , batch_size=_lowercase , seq_length=_lowercase , framework=_lowercase ) __a : Any = super().generate_dummy_inputs( processor.image_processor , batch_size=_lowercase , framework=_lowercase ) return {**text_input_dict, **image_input_dict} @property def lowerCAmelCase__(self ): '''simple docstring''' return 14
581
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __lowerCAmelCase ( __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Dict = checkpoints.load_tax_checkpoint(__UpperCamelCase ) snake_case_ : Tuple = flatten_dict(__UpperCamelCase ) return flax_params def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] ): '''simple docstring''' snake_case_ : Optional[Any] = {} snake_case_ : List[Any] = { """token_embedder""": """embeddings""", """encoder_norm""": """layernorm""", """kernel""": """weight""", """.out""": """.output""", """scale""": """weight""", """embedders_0.pos_embedding""": """row_embedder.weight""", """embedders_1.pos_embedding""": """column_embedder.weight""", } snake_case_ : Optional[Any] = { """query""": """attention.query""", """key""": """attention.key""", """value""": """attention.value""", """output.dense""": """output""", """encoder_decoder_attention.o""": """encoder_decoder_attention.attention.o""", """pre_self_attention_layer_norm""": """self_attention.layer_norm""", """pre_cross_attention_layer_norm""": """encoder_decoder_attention.layer_norm""", """mlp.""": """mlp.DenseReluDense.""", """pre_mlp_layer_norm""": """mlp.layer_norm""", """self_attention.o""": """self_attention.attention.o""", """decoder.embeddings.embedding""": """decoder.embed_tokens.weight""", """decoder.relpos_bias.rel_embedding""": """decoder.layer.0.self_attention.attention.relative_attention_bias.weight""", """decoder.decoder_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.logits_dense.weight""": """decoder.lm_head.weight""", } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key snake_case_ : List[Any] = """.""".join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): snake_case_ : List[str] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): snake_case_ : Optional[int] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number snake_case_ : Optional[Any] = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Union[str, Any] = new_key.replace("""encoder""" , """encoder.encoder""" ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number snake_case_ : int = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Dict = flax_dict[key] snake_case_ : Tuple = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): snake_case_ : Optional[int] = torch.from_numpy(converted_dict[key].T ) else: snake_case_ : List[Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : List[str]=False ): '''simple docstring''' snake_case_ : Optional[int] = get_flax_param(__UpperCamelCase ) if not use_large: snake_case_ : Optional[int] = PixaStructVisionConfig() snake_case_ : Optional[Any] = PixaStructTextConfig() else: snake_case_ : Tuple = PixaStructVisionConfig( hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_attention_heads=2_4 , num_hidden_layers=1_8 ) snake_case_ : List[str] = PixaStructTextConfig(hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_heads=2_4 , num_layers=1_8 ) snake_case_ : str = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=__UpperCamelCase ) snake_case_ : Optional[int] = PixaStructForConditionalGeneration(__UpperCamelCase ) snake_case_ : str = rename_and_convert_flax_params(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) snake_case_ : Optional[int] = AutoTokenizer.from_pretrained("""ybelkada/test-pix2struct-tokenizer""" ) snake_case_ : int = PixaStructImageProcessor() snake_case_ : str = PixaStructProcessor(image_processor=__UpperCamelCase , tokenizer=__UpperCamelCase ) if use_large: snake_case_ : Optional[Any] = 4_0_9_6 snake_case_ : int = True # mkdir if needed os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) processor.save_pretrained(__UpperCamelCase ) print("""Model saved in {}""".format(__UpperCamelCase ) ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') __lowerCAmelCase : List[Any] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
58
0
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.esm.modeling_esmfold import EsmForProteinFolding class _lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase=13 , UpperCAmelCase=7 , UpperCAmelCase=False , UpperCAmelCase=True , UpperCAmelCase=False , UpperCAmelCase=False , UpperCAmelCase=19 , UpperCAmelCase=32 , UpperCAmelCase=5 , UpperCAmelCase=4 , UpperCAmelCase=37 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=16 , UpperCAmelCase=2 , UpperCAmelCase=0.02 , UpperCAmelCase=3 , UpperCAmelCase=4 , UpperCAmelCase=None , ) -> Dict: '''simple docstring''' __snake_case : Any = parent __snake_case : str = batch_size __snake_case : Dict = seq_length __snake_case : List[str] = is_training __snake_case : Tuple = use_input_mask __snake_case : str = use_token_type_ids __snake_case : Any = use_labels __snake_case : Any = vocab_size __snake_case : Union[str, Any] = hidden_size __snake_case : Any = num_hidden_layers __snake_case : Dict = num_attention_heads __snake_case : Dict = intermediate_size __snake_case : List[str] = hidden_act __snake_case : Any = hidden_dropout_prob __snake_case : Optional[int] = attention_probs_dropout_prob __snake_case : List[Any] = max_position_embeddings __snake_case : Optional[Any] = type_vocab_size __snake_case : List[str] = type_sequence_label_size __snake_case : str = initializer_range __snake_case : List[str] = num_labels __snake_case : Optional[Any] = num_choices __snake_case : str = scope def UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' __snake_case : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : int = None if self.use_input_mask: __snake_case : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case : List[Any] = None __snake_case : Optional[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case : str = ids_tensor([self.batch_size] , self.num_choices ) __snake_case : str = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase ( self ) -> Any: '''simple docstring''' __snake_case : Dict = EsmConfig( vocab_size=33 , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , is_folding_model=_lowercase , esmfold_config={"trunk": {"num_blocks": 2}, "fp16_esm": False} , ) return config def UpperCAmelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' __snake_case : Optional[int] = EsmForProteinFolding(config=_lowercase ).float() model.to(_lowercase ) model.eval() __snake_case : Tuple = model(_lowercase , attention_mask=_lowercase ) __snake_case : int = model(_lowercase ) __snake_case : Any = model(_lowercase ) self.parent.assertEqual(result.positions.shape , (8, self.batch_size, self.seq_length, 14, 3) ) self.parent.assertEqual(result.angles.shape , (8, self.batch_size, self.seq_length, 7, 2) ) def UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' __snake_case : List[Any] = self.prepare_config_and_inputs() ( __snake_case ) : Dict = config_and_inputs __snake_case : str = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ : Optional[Any] =False UpperCAmelCase_ : int =(EsmForProteinFolding,) if is_torch_available() else () UpperCAmelCase_ : Optional[Any] =() UpperCAmelCase_ : int ={} if is_torch_available() else {} UpperCAmelCase_ : Any =False def UpperCAmelCase ( self ) -> Any: '''simple docstring''' __snake_case : str = EsmFoldModelTester(self ) __snake_case : Union[str, Any] = ConfigTester(self , config_class=_lowercase , hidden_size=37 ) def UpperCAmelCase ( self ) -> int: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) @unittest.skip("Does not support attention outputs" ) def UpperCAmelCase ( self ) -> Any: '''simple docstring''' pass @unittest.skip def UpperCAmelCase ( self ) -> Dict: '''simple docstring''' pass @unittest.skip("Esm does not support embedding resizing" ) def UpperCAmelCase ( self ) -> Dict: '''simple docstring''' pass @unittest.skip("Esm does not support embedding resizing" ) def UpperCAmelCase ( self ) -> str: '''simple docstring''' pass @unittest.skip("ESMFold does not support passing input embeds!" ) def UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' pass @unittest.skip("ESMFold does not support head pruning." ) def UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' pass @unittest.skip("ESMFold does not support head pruning." ) def UpperCAmelCase ( self ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip("ESMFold does not support head pruning." ) def UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' pass @unittest.skip("ESMFold does not support head pruning." ) def UpperCAmelCase ( self ) -> Any: '''simple docstring''' pass @unittest.skip("ESMFold does not support head pruning." ) def UpperCAmelCase ( self ) -> str: '''simple docstring''' pass @unittest.skip("ESMFold does not output hidden states in the normal way." ) def UpperCAmelCase ( self ) -> int: '''simple docstring''' pass @unittest.skip("ESMfold does not output hidden states in the normal way." ) def UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip("ESMFold only has one output format." ) def UpperCAmelCase ( self ) -> Dict: '''simple docstring''' pass @unittest.skip("This test doesn't work for ESMFold and doesn't test core functionality" ) def UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' pass @unittest.skip("ESMFold does not support input chunking." ) def UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' pass @unittest.skip("ESMFold doesn't respect you and it certainly doesn't respect your initialization arguments." ) def UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip("ESMFold doesn't support torchscript compilation." ) def UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' pass @unittest.skip("ESMFold doesn't support torchscript compilation." ) def UpperCAmelCase ( self ) -> int: '''simple docstring''' pass @unittest.skip("ESMFold doesn't support torchscript compilation." ) def UpperCAmelCase ( self ) -> Any: '''simple docstring''' pass @unittest.skip("ESMFold doesn't support data parallel." ) def UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' pass @require_torch class _lowerCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @slow def UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' __snake_case : List[str] = EsmForProteinFolding.from_pretrained("facebook/esmfold_v1" ).float() model.eval() __snake_case : Any = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) __snake_case : Tuple = model(_lowercase )["""positions"""] __snake_case : Any = torch.tensor([2.5_828, 0.7_993, -10.9_334] , dtype=torch.floataa ) self.assertTrue(torch.allclose(position_outputs[0, 0, 0, 0] , _lowercase , atol=1E-4 ) )
243
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if mass < 0: raise ValueError("""The mass of a body cannot be negative""" ) return 0.5 * mass * abs(__UpperCamelCase ) * abs(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
58
0
"""simple docstring""" from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging UpperCamelCase_ : Optional[int] = logging.get_logger(__name__) class __lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["pixel_values"] def __init__( self : int , _snake_case : Tuple = True , _snake_case : List[str] = 1 / 255 , _snake_case : Union[str, Any] = True , _snake_case : Optional[int] = 8 , **_snake_case : Any , ) -> None: """simple docstring""" super().__init__(**_lowercase ) A_ = do_rescale A_ = rescale_factor A_ = do_pad A_ = pad_size def lowerCamelCase__ ( self : str , _snake_case : Optional[Any] , _snake_case : str , _snake_case : int = None , **_snake_case : List[str] ) -> np.ndarray: """simple docstring""" return rescale(_lowercase , scale=_lowercase , data_format=_lowercase , **_lowercase ) def lowerCamelCase__ ( self : Optional[int] , _snake_case : List[Any] , _snake_case : Any , _snake_case : Any = None ) -> List[Any]: """simple docstring""" A_ = get_image_size(_lowercase ) A_ = (old_height // size + 1) * size - old_height A_ = (old_width // size + 1) * size - old_width return pad(_lowercase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=_lowercase ) def lowerCamelCase__ ( self : str , _snake_case : int , _snake_case : Optional[Any] = None , _snake_case : Tuple = None , _snake_case : List[str] = None , _snake_case : str = None , _snake_case : Dict = None , _snake_case : int = ChannelDimension.FIRST , **_snake_case : Tuple , ) -> Any: """simple docstring""" A_ = do_rescale if do_rescale is not None else self.do_rescale A_ = rescale_factor if rescale_factor is not None else self.rescale_factor A_ = do_pad if do_pad is not None else self.do_pad A_ = pad_size if pad_size is not None else self.pad_size A_ = make_list_of_images(_lowercase ) if not valid_images(_lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. A_ = [to_numpy_array(_lowercase ) for image in images] if do_rescale: A_ = [self.rescale(image=_lowercase , scale=_lowercase ) for image in images] if do_pad: A_ = [self.pad(_lowercase , size=_lowercase ) for image in images] A_ = [to_channel_dimension_format(_lowercase , _lowercase ) for image in images] A_ = {"""pixel_values""": images} return BatchFeature(data=_lowercase , tensor_type=_lowercase )
115
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = StableDiffusionInpaintPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase = frozenset([] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=_lowercase , ) snake_case_ : Dict = PNDMScheduler(skip_prk_steps=_lowercase ) torch.manual_seed(0 ) snake_case_ : str = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) snake_case_ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) snake_case_ : Dict = CLIPTextModel(_lowercase ) snake_case_ : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case_ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self , _lowercase , _lowercase=0 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowercase ) ).to(_lowercase ) snake_case_ : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case_ : Tuple = Image.fromarray(np.uinta(_lowercase ) ).convert("""RGB""" ).resize((6_4, 6_4) ) snake_case_ : Any = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((6_4, 6_4) ) if str(_lowercase ).startswith("""mps""" ): snake_case_ : str = torch.manual_seed(_lowercase ) else: snake_case_ : List[str] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) snake_case_ : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case_ : List[str] = self.get_dummy_components() snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline(**_lowercase ) snake_case_ : Dict = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) snake_case_ : Optional[int] = self.get_dummy_inputs(_lowercase ) snake_case_ : List[str] = sd_pipe(**_lowercase ).images snake_case_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) snake_case_ : Optional[int] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained(_lowercase , safety_checker=_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : List[Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[str] = torch.manual_seed(0 ) snake_case_ : Dict = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ : str = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , torch_dtype=torch.floataa , safety_checker=_lowercase , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : Union[str, Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : Optional[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : str = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case_ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : int = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : List[str] = PNDMScheduler.from_pretrained(_lowercase , subfolder="""scheduler""" ) snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , safety_checker=_lowercase , scheduler=_lowercase , torch_dtype=torch.floataa , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case_ : Any = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , num_inference_steps=2 , output_type="""np""" , ) snake_case_ : Dict = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
58
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCAmelCase_ = '''roformer''' def __init__( self : int , _A : Dict=5_0000 , _A : Union[str, Any]=None , _A : Optional[int]=768 , _A : Dict=12 , _A : List[str]=12 , _A : Union[str, Any]=3072 , _A : List[Any]="gelu" , _A : int=0.1 , _A : Any=0.1 , _A : int=1536 , _A : List[Any]=2 , _A : Optional[Any]=0.02 , _A : List[str]=1e-12 , _A : Any=0 , _A : Any=False , _A : List[str]=True , **_A : Optional[Any] , ): """simple docstring""" super().__init__(pad_token_id=_lowercase , **_lowercase ) __SCREAMING_SNAKE_CASE : str = vocab_size __SCREAMING_SNAKE_CASE : Any = hidden_size if embedding_size is None else embedding_size __SCREAMING_SNAKE_CASE : List[str] = hidden_size __SCREAMING_SNAKE_CASE : str = num_hidden_layers __SCREAMING_SNAKE_CASE : Union[str, Any] = num_attention_heads __SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_act __SCREAMING_SNAKE_CASE : Optional[Any] = intermediate_size __SCREAMING_SNAKE_CASE : List[str] = hidden_dropout_prob __SCREAMING_SNAKE_CASE : str = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE : Union[str, Any] = max_position_embeddings __SCREAMING_SNAKE_CASE : List[str] = type_vocab_size __SCREAMING_SNAKE_CASE : Tuple = initializer_range __SCREAMING_SNAKE_CASE : str = layer_norm_eps __SCREAMING_SNAKE_CASE : List[str] = rotary_value __SCREAMING_SNAKE_CASE : str = use_cache class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self : Any ): """simple docstring""" if self.task == "multiple-choice": __SCREAMING_SNAKE_CASE : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __SCREAMING_SNAKE_CASE : Any = {0: """batch""", 1: """sequence"""} __SCREAMING_SNAKE_CASE : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
74
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : str , __UpperCamelCase : Any ): '''simple docstring''' snake_case_ : Optional[Any] = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] snake_case_ : Optional[int] = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } snake_case_ : Optional[Any] = F'{src_lang}-{tgt_lang}' snake_case_ : Dict = F'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) snake_case_ : List[str] = os.path.join(__UpperCamelCase , """README.md""" ) print(F'Generating {path}' ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(__UpperCamelCase ) # make sure we are under the root of the project __lowerCAmelCase : str = Path(__file__).resolve().parent.parent.parent __lowerCAmelCase : Optional[int] = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[int] = model_name.split('''-''') __lowerCAmelCase : Optional[int] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
58
0
'''simple docstring''' class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): pass class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): pass class __UpperCamelCase : def __init__( self ): """simple docstring""" lowerCamelCase_ =[ [], [], [], ] def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" try: if len(self.queues[priority] ) >= 100: raise OverflowError('''Maximum queue size is 100''' ) self.queues[priority].append(_lowercase ) except IndexError: raise ValueError('''Valid priorities are 0, 1, and 2''' ) def lowercase__ ( self ): """simple docstring""" for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError('''All queues are empty''' ) def __str__( self ): """simple docstring""" return "\n".join(f'''Priority {i}: {q}''' for i, q in enumerate(self.queues ) ) class __UpperCamelCase : def __init__( self ): """simple docstring""" lowerCamelCase_ =[] def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" if len(self.queue ) == 100: raise OverFlowError('''Maximum queue size is 100''' ) self.queue.append(_lowercase ) def lowercase__ ( self ): """simple docstring""" if not self.queue: raise UnderFlowError('''The queue is empty''' ) else: lowerCamelCase_ =min(self.queue ) self.queue.remove(_lowercase ) return data def __str__( self ): """simple docstring""" return str(self.queue ) def a_ ( ) -> List[Any]: """simple docstring""" lowerCamelCase_ =FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(__UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(__UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def a_ ( ) -> Any: """simple docstring""" lowerCamelCase_ =ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(__UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(__UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
676
"""simple docstring""" __lowerCAmelCase : Tuple = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowerCAmelCase : Tuple = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __lowerCAmelCase : Any = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
58
0
import requests _lowerCAmelCase : int ='''https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=''' def _A ( SCREAMING_SNAKE_CASE ): UpperCAmelCase__: Dict = requests.get(_NEWS_API + bbc_news_api_key ).json() # each article in the list is a dict for i, article in enumerate(bbc_news_page["articles"] ,1 ): print(f"{i}.) {article['title']}" ) if __name__ == "__main__": fetch_bbc_news(bbc_news_api_key="""<Your BBC News API key goes here>""")
113
"""simple docstring""" from jiwer import compute_measures import datasets __lowerCAmelCase : Tuple = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __lowerCAmelCase : Union[str, Any] = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' __lowerCAmelCase : Optional[int] = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", ] , ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=False ) -> Optional[Any]: '''simple docstring''' if concatenate_texts: return compute_measures(_lowercase , _lowercase )["wer"] else: snake_case_ : List[str] = 0 snake_case_ : Optional[int] = 0 for prediction, reference in zip(_lowercase , _lowercase ): snake_case_ : Optional[Any] = compute_measures(_lowercase , _lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
58
0
"""simple docstring""" import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin __UpperCamelCase : List[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __UpperCamelCase : Tuple = 2_5_6_0_4_7 __UpperCamelCase : int = 2_5_6_1_4_5 @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" lowercase__ = NllbTokenizer lowercase__ = NllbTokenizerFast lowercase__ = True lowercase__ = True lowercase__ = {} def __lowerCAmelCase ( self : Optional[Any] ): super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase__ : str = NllbTokenizer(_lowercase ,keep_accents=_lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self : Dict ): lowerCAmelCase__ : Tuple = NllbTokenizer(_lowercase ,keep_accents=_lowercase ) lowerCAmelCase__ : Optional[Any] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowercase ,['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowercase ) ,[value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] ,) lowerCAmelCase__ : Optional[int] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowercase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] ,) lowerCAmelCase__ : Union[str, Any] = tokenizer.convert_tokens_to_ids(_lowercase ) self.assertListEqual( _lowercase ,[ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] ,) lowerCAmelCase__ : List[Any] = tokenizer.convert_ids_to_tokens(_lowercase ) self.assertListEqual( _lowercase ,[ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] ,) def __lowerCAmelCase ( self : Union[str, Any] ): lowerCAmelCase__ : Tuple = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-nllb""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_lowercase ,**_lowercase ) lowerCAmelCase__ : Optional[int] = self.tokenizer_class.from_pretrained(_lowercase ,**_lowercase ) lowerCAmelCase__ : Any = tempfile.mkdtemp() lowerCAmelCase__ : Optional[int] = tokenizer_r.save_pretrained(_lowercase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.save_pretrained(_lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) lowerCAmelCase__ : Dict = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(_lowercase ,_lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase__ : Optional[int] = tokenizer_r.from_pretrained(_lowercase ) lowerCAmelCase__ : str = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase ,_lowercase ) ) shutil.rmtree(_lowercase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase__ : Optional[Any] = tempfile.mkdtemp() lowerCAmelCase__ : Optional[Any] = tokenizer_r.save_pretrained(_lowercase ,legacy_format=_lowercase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.save_pretrained(_lowercase ) # Checks it save with the same files self.assertSequenceEqual(_lowercase ,_lowercase ) # Checks everything loads correctly in the same way lowerCAmelCase__ : Union[str, Any] = tokenizer_r.from_pretrained(_lowercase ) lowerCAmelCase__ : str = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase ,_lowercase ) ) shutil.rmtree(_lowercase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase__ : str = tempfile.mkdtemp() lowerCAmelCase__ : Any = tokenizer_r.save_pretrained(_lowercase ,legacy_format=_lowercase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.save_pretrained(_lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase__ : List[str] = tokenizer_r.from_pretrained(_lowercase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase ,_lowercase ) ) shutil.rmtree(_lowercase ) @require_torch def __lowerCAmelCase ( self : List[str] ): if not self.test_seqaseq: return lowerCAmelCase__ : Union[str, Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # Longer text that will definitely require truncation. lowerCAmelCase__ : Optional[Any] = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for""" """ Syria is that 'there is no military solution' to the nearly five-year conflict and more weapons""" """ will only worsen the violence and misery for millions of people.""", ] lowerCAmelCase__ : Optional[int] = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al""" """ Rusiei pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi""" """ că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] try: lowerCAmelCase__ : Any = tokenizer.prepare_seqaseq_batch( src_texts=_lowercase ,tgt_texts=_lowercase ,max_length=3 ,max_target_length=1_0 ,return_tensors='''pt''' ,src_lang='''eng_Latn''' ,tgt_lang='''ron_Latn''' ,) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] ,3 ) self.assertEqual(batch.labels.shape[1] ,1_0 ) # max_target_length will default to max_length if not specified lowerCAmelCase__ : str = tokenizer.prepare_seqaseq_batch( _lowercase ,tgt_texts=_lowercase ,max_length=3 ,return_tensors='''pt''' ) self.assertEqual(batch.input_ids.shape[1] ,3 ) self.assertEqual(batch.labels.shape[1] ,3 ) lowerCAmelCase__ : Optional[int] = tokenizer.prepare_seqaseq_batch( src_texts=_lowercase ,max_length=3 ,max_target_length=1_0 ,return_tensors='''pt''' ) self.assertEqual(batch_encoder_only.input_ids.shape[1] ,3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] ,3 ) self.assertNotIn('''decoder_input_ids''' ,_lowercase ) @unittest.skip('''Unfortunately way too slow to build a BPE with SentencePiece.''' ) def __lowerCAmelCase ( self : Tuple ): pass def __lowerCAmelCase ( self : Dict ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): lowerCAmelCase__ : List[str] = [AddedToken('''<special>''' ,lstrip=_lowercase )] lowerCAmelCase__ : List[Any] = self.rust_tokenizer_class.from_pretrained( _lowercase ,additional_special_tokens=_lowercase ,**_lowercase ) lowerCAmelCase__ : Dict = tokenizer_r.encode('''Hey this is a <special> token''' ) lowerCAmelCase__ : int = tokenizer_r.encode('''<special>''' ,add_special_tokens=_lowercase )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: lowerCAmelCase__ : Optional[Any] = self.rust_tokenizer_class.from_pretrained( _lowercase ,additional_special_tokens=_lowercase ,**_lowercase ,) lowerCAmelCase__ : List[Any] = self.tokenizer_class.from_pretrained( _lowercase ,additional_special_tokens=_lowercase ,**_lowercase ) lowerCAmelCase__ : Tuple = tokenizer_p.encode('''Hey this is a <special> token''' ) lowerCAmelCase__ : int = tokenizer_cr.encode('''Hey this is a <special> token''' ) self.assertEqual(_lowercase ,_lowercase ) self.assertEqual(_lowercase ,_lowercase ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" lowercase__ = "facebook/nllb-200-distilled-600M" lowercase__ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase__ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase__ = [ 256047, 16297, 134408, 8165, 248066, 14734, 950, 1135, 105721, 3573, 83, 27352, 108, 49486, 2, ] @classmethod def __lowerCAmelCase ( cls : Optional[Any] ): lowerCAmelCase__ : NllbTokenizer = NllbTokenizer.from_pretrained( cls.checkpoint_name ,src_lang='''eng_Latn''' ,tgt_lang='''ron_Latn''' ) lowerCAmelCase__ : Dict = 1 return cls def __lowerCAmelCase ( self : Dict ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ace_Arab'''] ,2_5_6_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ace_Latn'''] ,2_5_6_0_0_2 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''fra_Latn'''] ,2_5_6_0_5_7 ) def __lowerCAmelCase ( self : Dict ): lowerCAmelCase__ : int = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens ,_lowercase ) def __lowerCAmelCase ( self : Dict ): self.assertIn(_lowercase ,self.tokenizer.all_special_ids ) # fmt: off lowerCAmelCase__ : List[Any] = [RO_CODE, 4_2_5_4, 9_8_0_6_8, 1_1_2_9_2_3, 3_9_0_7_2, 3_9_0_9, 7_1_3, 1_0_2_7_6_7, 2_6, 1_7_3_1_4, 3_5_6_4_2, 1_4_6_8_3, 3_3_1_1_8, 2_0_2_2, 6_6_9_8_7, 2, 2_5_6_0_4_7] # fmt: on lowerCAmelCase__ : int = self.tokenizer.decode(_lowercase ,skip_special_tokens=_lowercase ) lowerCAmelCase__ : List[str] = self.tokenizer.decode(generated_ids[1:] ,skip_special_tokens=_lowercase ) self.assertEqual(_lowercase ,_lowercase ) self.assertNotIn(self.tokenizer.eos_token ,_lowercase ) def __lowerCAmelCase ( self : Optional[int] ): lowerCAmelCase__ : List[str] = ["""this is gunna be a long sentence """ * 2_0] assert isinstance(src_text[0] ,_lowercase ) lowerCAmelCase__ : Optional[int] = 1_0 lowerCAmelCase__ : Optional[int] = self.tokenizer(_lowercase ,max_length=_lowercase ,truncation=_lowercase ).input_ids[0] self.assertEqual(ids[-1] ,2 ) self.assertEqual(ids[0] ,_lowercase ) self.assertEqual(len(_lowercase ) ,_lowercase ) def __lowerCAmelCase ( self : List[str] ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) ,[2_5_6_2_0_3, 3] ) def __lowerCAmelCase ( self : List[str] ): lowerCAmelCase__ : Dict = tempfile.mkdtemp() lowerCAmelCase__ : List[Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_lowercase ) lowerCAmelCase__ : Union[str, Any] = NllbTokenizer.from_pretrained(_lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids ,_lowercase ) @require_torch def __lowerCAmelCase ( self : Union[str, Any] ): lowerCAmelCase__ : str = self.tokenizer( self.src_text ,text_target=self.tgt_text ,padding=_lowercase ,truncation=_lowercase ,max_length=len(self.expected_src_tokens ) ,return_tensors='''pt''' ,) lowerCAmelCase__ : Optional[Any] = shift_tokens_right( batch['''labels'''] ,self.tokenizer.pad_token_id ,self.tokenizer.lang_code_to_id['''ron_Latn'''] ) self.assertIsInstance(_lowercase ,_lowercase ) self.assertEqual((2, 1_5) ,batch.input_ids.shape ) self.assertEqual((2, 1_5) ,batch.attention_mask.shape ) lowerCAmelCase__ : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens ,_lowercase ) self.assertEqual(_lowercase ,batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens ,[EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens ,[self.tokenizer.eos_token_id] ) def __lowerCAmelCase ( self : List[Any] ): lowerCAmelCase__ : Tuple = self.tokenizer(self.src_text ,padding=_lowercase ,truncation=_lowercase ,max_length=3 ,return_tensors='''pt''' ) lowerCAmelCase__ : Optional[Any] = self.tokenizer( text_target=self.tgt_text ,padding=_lowercase ,truncation=_lowercase ,max_length=1_0 ,return_tensors='''pt''' ) lowerCAmelCase__ : str = targets["""input_ids"""] lowerCAmelCase__ : Optional[int] = shift_tokens_right( _lowercase ,self.tokenizer.pad_token_id ,decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] ,) self.assertEqual(batch.input_ids.shape[1] ,3 ) self.assertEqual(batch.decoder_input_ids.shape[1] ,1_0 ) @require_torch def __lowerCAmelCase ( self : str ): lowerCAmelCase__ : List[str] = self.tokenizer._build_translation_inputs( '''A test''' ,return_tensors='''pt''' ,src_lang='''eng_Latn''' ,tgt_lang='''fra_Latn''' ) self.assertEqual( nested_simplify(_lowercase ) ,{ # A, test, EOS, en_XX '''input_ids''': [[2_5_6_0_4_7, 7_0, 7_3_5_6, 2]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 2_5_6_0_5_7, } ,) @require_torch def __lowerCAmelCase ( self : Dict ): lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Optional[int] = self.tokenizer( '''UN Chief says there is no military solution in Syria''' ,src_lang='''eng_Latn''' ,tgt_lang='''fra_Latn''' ) self.assertEqual( inputs.input_ids ,[1_6_2_9_7, 1_3_4_4_0_8, 2_5_6_5_3, 6_3_7_0, 2_4_8, 2_5_4, 1_0_3_9_2_9, 9_4_9_9_5, 1_0_8, 4_9_4_8_6, 2, 2_5_6_0_4_7] ) lowerCAmelCase__ : str = False lowerCAmelCase__ : List[Any] = self.tokenizer( '''UN Chief says there is no military solution in Syria''' ,src_lang='''eng_Latn''' ,tgt_lang='''fra_Latn''' ) self.assertEqual( inputs.input_ids ,[2_5_6_0_4_7, 1_6_2_9_7, 1_3_4_4_0_8, 2_5_6_5_3, 6_3_7_0, 2_4_8, 2_5_4, 1_0_3_9_2_9, 9_4_9_9_5, 1_0_8, 4_9_4_8_6, 2] )
450
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1_3 , _lowercase=3 , _lowercase=2_2_4 , _lowercase=3_0 , _lowercase=4_0_0 , _lowercase=True , _lowercase=None , _lowercase=True , _lowercase=[0.5, 0.5, 0.5] , _lowercase=[0.5, 0.5, 0.5] , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : Optional[Any] = batch_size snake_case_ : Dict = num_channels snake_case_ : Optional[Any] = image_size snake_case_ : Optional[Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[int] = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : Dict = image_std def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ViTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = EfficientFormerImageProcessorTester(self ) @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.image_proc_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , """image_mean""" ) ) self.assertTrue(hasattr(_lowercase , """image_std""" ) ) self.assertTrue(hasattr(_lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowercase , """do_resize""" ) ) self.assertTrue(hasattr(_lowercase , """size""" ) ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' pass def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input snake_case_ : Any = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Optional[Any] = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Dict = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input snake_case_ : int = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : int = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input snake_case_ : Dict = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Tuple = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
58
0
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class a_ ( unittest.TestCase , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ = load_tool('text-classification' ) self.tool.setup() lowerCAmelCase_ = load_tool('text-classification' , remote=_lowercase ) def _lowercase ( self ) -> List[str]: '''simple docstring''' lowerCAmelCase_ = self.tool('That\'s quite cool' , ['positive', 'negative'] ) self.assertEqual(_lowercase , 'positive' ) def _lowercase ( self ) -> int: '''simple docstring''' lowerCAmelCase_ = self.remote_tool('That\'s quite cool' , ['positive', 'negative'] ) self.assertEqual(_lowercase , 'positive' ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ = self.tool(text='That\'s quite cool' , labels=['positive', 'negative'] ) self.assertEqual(_lowercase , 'positive' ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ = self.remote_tool(text='That\'s quite cool' , labels=['positive', 'negative'] ) self.assertEqual(_lowercase , 'positive' )
318
"""simple docstring""" from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCAmelCase : int = TypeVar('''KT''') __lowerCAmelCase : Union[str, Any] = TypeVar('''VT''') class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = "root" , _lowercase = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = key snake_case_ : Tuple = value snake_case_ : list[Node[KT, VT]] = [] def __repr__( self ) -> str: '''simple docstring''' return f'Node({self.key}: {self.value})' @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return len(self.forward ) class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = 0.5 , _lowercase = 1_6 ) -> int: '''simple docstring''' snake_case_ : Node[KT, VT] = Node[KT, VT]() snake_case_ : Union[str, Any] = 0 snake_case_ : Optional[int] = p snake_case_ : Any = max_level def __str__( self ) -> str: '''simple docstring''' snake_case_ : str = list(self ) if len(_lowercase ) == 0: return f'SkipList(level={self.level})' snake_case_ : List[Any] = max((len(str(_lowercase ) ) for item in items) , default=4 ) snake_case_ : str = max(_lowercase , 4 ) + 4 snake_case_ : Union[str, Any] = self.head snake_case_ : Dict = [] snake_case_ : List[str] = node.forward.copy() lines.append(f'[{node.key}]'.ljust(_lowercase , """-""" ) + """* """ * len(_lowercase ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) while len(node.forward ) != 0: snake_case_ : Optional[Any] = node.forward[0] lines.append( f'[{node.key}]'.ljust(_lowercase , """-""" ) + """ """.join(str(n.key ) if n.key == node.key else """|""" for n in forwards ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) snake_case_ : List[str] = node.forward lines.append("""None""".ljust(_lowercase ) + """* """ * len(_lowercase ) ) return f'SkipList(level={self.level})\n' + "\n".join(_lowercase ) def __iter__( self ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.head while len(node.forward ) != 0: yield node.forward[0].key snake_case_ : Dict = node.forward[0] def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[int] = 1 while random() < self.p and level < self.max_level: level += 1 return level def UpperCAmelCase__ ( self , _lowercase ) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: snake_case_ : List[Any] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(_lowercase ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: for i, update_node in enumerate(_lowercase ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: snake_case_ : List[str] = node.forward[i] else: snake_case_ : Tuple = update_node.forward[:i] def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> str: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: snake_case_ : List[Any] = value else: snake_case_ : Optional[int] = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , _lowercase ): update_vector.append(self.head ) snake_case_ : Any = level snake_case_ : Optional[int] = Node(_lowercase , _lowercase ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(_lowercase ) else: snake_case_ : Optional[Any] = new_node def UpperCAmelCase__ ( self , _lowercase ) -> VT | None: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: return node.value return None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : List[str] = SkipList() skip_list.insert("""Key1""" , 3 ) skip_list.insert("""Key2""" , 1_2 ) skip_list.insert("""Key3""" , 4_1 ) skip_list.insert("""Key4""" , -1_9 ) snake_case_ : Optional[int] = skip_list.head snake_case_ : List[Any] = {} while node.level != 0: snake_case_ : List[str] = node.forward[0] snake_case_ : Union[str, Any] = node.value assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 1_2 assert all_values["Key3"] == 4_1 assert all_values["Key4"] == -1_9 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_0 ) skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""Key5""" , 7 ) skip_list.insert("""Key7""" , 1_0 ) skip_list.insert("""Key10""" , 5 ) skip_list.insert("""Key7""" , 7 ) skip_list.insert("""Key5""" , 5 ) skip_list.insert("""Key10""" , 1_0 ) snake_case_ : str = skip_list.head snake_case_ : str = {} while node.level != 0: snake_case_ : Optional[Any] = node.forward[0] snake_case_ : int = node.value if len(__UpperCamelCase ) != 4: print() assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 1_2 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 1_0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : str = SkipList() assert skip_list.find("""Some key""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[Any] = SkipList() skip_list.insert("""Key2""" , 2_0 ) assert skip_list.find("""Key2""" ) == 2_0 skip_list.insert("""Some Key""" , 1_0 ) skip_list.insert("""Key2""" , 8 ) skip_list.insert("""V""" , 1_3 ) assert skip_list.find("""Y""" ) is None assert skip_list.find("""Key2""" ) == 8 assert skip_list.find("""Some Key""" ) == 1_0 assert skip_list.find("""V""" ) == 1_3 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Any = SkipList() skip_list.delete("""Some key""" ) assert len(skip_list.head.forward ) == 0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Tuple = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) == 1_4 assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""X""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key1""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Union[str, Any] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4_2 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""X""" ) def traverse_keys(__UpperCamelCase : str ): yield node.key for forward_node in node.forward: yield from traverse_keys(__UpperCamelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __lowerCAmelCase ( ): '''simple docstring''' def is_sorted(__UpperCamelCase : List[Any] ): return all(next_item >= item for item, next_item in zip(__UpperCamelCase , lst[1:] ) ) snake_case_ : str = SkipList() for i in range(1_0 ): skip_list.insert(__UpperCamelCase , __UpperCamelCase ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.insert(-1_2 , -1_2 ) skip_list.insert(7_7 , 7_7 ) assert is_sorted(list(__UpperCamelCase ) ) def __lowerCAmelCase ( ): '''simple docstring''' for _ in range(1_0_0 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = SkipList() skip_list.insert(2 , """2""" ) skip_list.insert(4 , """4""" ) skip_list.insert(6 , """4""" ) skip_list.insert(4 , """5""" ) skip_list.insert(8 , """4""" ) skip_list.insert(9 , """4""" ) skip_list.delete(4 ) print(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
58
0
"""simple docstring""" import math from collections.abc import Iterator from itertools import takewhile def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5, int(math.sqrt(__UpperCamelCase ) + 1 ), 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCamelCase__ ( ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = 2 while True: if is_prime(__UpperCamelCase ): yield num num += 1 def lowerCamelCase__ ( __snake_case = 2_00_00_00 ) -> List[Any]: """simple docstring""" return sum(takewhile(lambda __snake_case : x < n, prime_generator() ) ) if __name__ == "__main__": print(F"""{solution() = }""")
19
"""simple docstring""" import argparse import os import re import packaging.version __lowerCAmelCase : Optional[Any] = '''examples/''' __lowerCAmelCase : Union[str, Any] = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __lowerCAmelCase : Union[str, Any] = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } __lowerCAmelCase : List[Any] = '''README.md''' def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : Dict ): '''simple docstring''' with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : Any = f.read() snake_case_ , snake_case_ : Optional[int] = REPLACE_PATTERNS[pattern] snake_case_ : Union[str, Any] = replace.replace("""VERSION""" , __UpperCamelCase ) snake_case_ : List[Any] = re_pattern.sub(__UpperCamelCase , __UpperCamelCase ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' for folder, directories, fnames in os.walk(__UpperCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(__UpperCamelCase , __UpperCamelCase ) , __UpperCamelCase , pattern="""examples""" ) def __lowerCAmelCase ( __UpperCamelCase : List[str] , __UpperCamelCase : int=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not patch: update_version_in_examples(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = """🤗 Transformers currently provides the following architectures""" snake_case_ : Union[str, Any] = """1. Want to contribute a new model?""" with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : str = f.readlines() # Find the start of the list. snake_case_ : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 snake_case_ : Optional[int] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): snake_case_ : Any = lines[index].replace( """https://huggingface.co/docs/diffusers/main/model_doc""" , """https://huggingface.co/docs/diffusers/model_doc""" , ) index += 1 with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] , """r""" ) as f: snake_case_ : Any = f.read() snake_case_ : Tuple = REPLACE_PATTERNS["""init"""][0].search(__UpperCamelCase ).groups()[0] return packaging.version.parse(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : str=False ): '''simple docstring''' snake_case_ : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: snake_case_ : str = default_version.base_version elif patch: snake_case_ : str = F'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: snake_case_ : str = F'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. snake_case_ : int = input(F'Which version are you releasing? [{default_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Optional[int] = default_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase , patch=__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = get_version() snake_case_ : str = F'{current_version.major}.{current_version.minor + 1}.0.dev0' snake_case_ : Tuple = current_version.base_version # Check with the user we got that right. snake_case_ : Optional[int] = input(F'Which version are we developing now? [{dev_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Dict = dev_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": __lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __lowerCAmelCase : str = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
58
0
'''simple docstring''' import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Dict=13 , __SCREAMING_SNAKE_CASE : Dict=7 , __SCREAMING_SNAKE_CASE : int=True , __SCREAMING_SNAKE_CASE : Any=True , __SCREAMING_SNAKE_CASE : Union[str, Any]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : List[str]=99 , __SCREAMING_SNAKE_CASE : int=32 , __SCREAMING_SNAKE_CASE : List[Any]=5 , __SCREAMING_SNAKE_CASE : int=4 , __SCREAMING_SNAKE_CASE : List[str]=37 , __SCREAMING_SNAKE_CASE : Union[str, Any]="gelu" , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : str=0.1 , __SCREAMING_SNAKE_CASE : Optional[int]=512 , __SCREAMING_SNAKE_CASE : Any=16 , __SCREAMING_SNAKE_CASE : Any=2 , __SCREAMING_SNAKE_CASE : Dict=0.02 , __SCREAMING_SNAKE_CASE : Optional[int]=4 , ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_attention_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_choices def UpperCAmelCase__ ( self : Tuple ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = None if self.use_attention_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase__ ( self : List[str] ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def UpperCAmelCase__ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class lowerCAmelCase__ ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = True lowerCAmelCase__ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase__ ( self : int ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRobertaPreLayerNormModelTester(self ) @slow def UpperCAmelCase__ ( self : str ) -> Optional[int]: """simple docstring""" for model_class_name in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""andreasmadsen/efficient_mlm_m0.40""" , from_pt=_lowercase ) __SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(_lowercase ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self : Dict ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("""andreasmadsen/efficient_mlm_m0.40""" , from_pt=_lowercase ) __SCREAMING_SNAKE_CASE = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __SCREAMING_SNAKE_CASE = model(_lowercase )[0] __SCREAMING_SNAKE_CASE = [1, 11, 50_265] self.assertEqual(list(output.shape ) , _lowercase ) # compare the actual values for a slice. __SCREAMING_SNAKE_CASE = np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _lowercase , atol=1E-4 ) ) @slow def UpperCAmelCase__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRobertaPreLayerNormModel.from_pretrained("""andreasmadsen/efficient_mlm_m0.40""" , from_pt=_lowercase ) __SCREAMING_SNAKE_CASE = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) __SCREAMING_SNAKE_CASE = model(_lowercase )[0] # compare the actual values for a slice. __SCREAMING_SNAKE_CASE = np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , _lowercase , atol=1E-4 ) )
627
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
58
0
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float = 1 / sqrt(2 ) ) -> List[str]: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = (1 - _cos) / 2 __snake_case = 1 - _cos __snake_case = 1 + alpha __snake_case = -2 * _cos __snake_case = 1 - alpha __snake_case = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float = 1 / sqrt(2 ) ) -> Tuple: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = (1 + _cos) / 2 __snake_case = -1 - _cos __snake_case = 1 + alpha __snake_case = -2 * _cos __snake_case = 1 - alpha __snake_case = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float = 1 / sqrt(2 ) ) -> str: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = _sin / 2 __snake_case = 0 __snake_case = -ba __snake_case = 1 + alpha __snake_case = -2 * _cos __snake_case = 1 - alpha __snake_case = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float = 1 / sqrt(2 ) ) -> Optional[Any]: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = 1 - alpha __snake_case = -2 * _cos __snake_case = 1 + alpha __snake_case = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float , snake_case_ : float = 1 / sqrt(2 ) , ) -> str: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = 10 ** (gain_db / 40) __snake_case = 1 + alpha * big_a __snake_case = -2 * _cos __snake_case = 1 - alpha * big_a __snake_case = 1 + alpha / big_a __snake_case = -2 * _cos __snake_case = 1 - alpha / big_a __snake_case = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float , snake_case_ : float = 1 / sqrt(2 ) , ) -> Union[str, Any]: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = 10 ** (gain_db / 40) __snake_case = (big_a + 1) - (big_a - 1) * _cos __snake_case = (big_a + 1) + (big_a - 1) * _cos __snake_case = (big_a - 1) - (big_a + 1) * _cos __snake_case = (big_a - 1) + (big_a + 1) * _cos __snake_case = 2 * sqrt(__UpperCamelCase ) * alpha __snake_case = big_a * (pmc + aaa) __snake_case = 2 * big_a * mpc __snake_case = big_a * (pmc - aaa) __snake_case = ppmc + aaa __snake_case = -2 * pmpc __snake_case = ppmc - aaa __snake_case = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int , snake_case_ : float , snake_case_ : float = 1 / sqrt(2 ) , ) -> List[Any]: __snake_case = tau * frequency / samplerate __snake_case = sin(__UpperCamelCase ) __snake_case = cos(__UpperCamelCase ) __snake_case = _sin / (2 * q_factor) __snake_case = 10 ** (gain_db / 40) __snake_case = (big_a + 1) - (big_a - 1) * _cos __snake_case = (big_a + 1) + (big_a - 1) * _cos __snake_case = (big_a - 1) - (big_a + 1) * _cos __snake_case = (big_a - 1) + (big_a + 1) * _cos __snake_case = 2 * sqrt(__UpperCamelCase ) * alpha __snake_case = big_a * (ppmc + aaa) __snake_case = -2 * big_a * pmpc __snake_case = big_a * (ppmc - aaa) __snake_case = pmc + aaa __snake_case = 2 * mpc __snake_case = pmc - aaa __snake_case = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
592
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) snake_case_ : str = precision snake_case_ : Any = ceil(precision / 1_4 ) snake_case_ : Dict = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : Optional[Any] = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : Optional[int] = Decimal(__UpperCamelCase ) for k in range(1 , __UpperCamelCase ): snake_case_ : Any = factorial(6 * k ) // (factorial(3 * k ) * factorial(__UpperCamelCase ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __lowerCAmelCase : int = 50 print(F'''The first {n} digits of pi is: {pi(n)}''')
58
0
"""simple docstring""" import random def __magic_name__ ( _lowerCamelCase : int , _lowerCamelCase : float , _lowerCamelCase : bool = False ): __a : dict = {i: [] for i in range(__UpperCamelCase )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(__UpperCamelCase ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(__UpperCamelCase ): for j in range(i + 1 , __UpperCamelCase ): if random.random() < probability: graph[i].append(__UpperCamelCase ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(__UpperCamelCase ) return graph def __magic_name__ ( _lowerCamelCase : int ): return { i: [j for j in range(__UpperCamelCase ) if i != j] for i in range(__UpperCamelCase ) } if __name__ == "__main__": import doctest doctest.testmod()
581
"""simple docstring""" import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCAmelCase ( __UpperCamelCase : Optional[int] ): '''simple docstring''' snake_case_ : Any = torch.exp(__UpperCamelCase ) snake_case_ : Optional[int] = torch.sum(__UpperCamelCase , dim=1 ) # sum of exp(x_i) snake_case_ : str = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(__UpperCamelCase ) - B / A class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> int: '''simple docstring''' super().__init__() snake_case_ : Tuple = config.output_attentions snake_case_ : str = config.output_hidden_states snake_case_ : List[str] = nn.ModuleList([BertLayer(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Tuple = nn.ModuleList([BertHighway(_lowercase ) for _ in range(config.num_hidden_layers )] ) snake_case_ : Any = [-1 for _ in range(config.num_hidden_layers )] def UpperCAmelCase__ ( self , _lowercase ) -> Tuple: '''simple docstring''' if (type(_lowercase ) is float) or (type(_lowercase ) is int): for i in range(len(self.early_exit_entropy ) ): snake_case_ : Dict = x else: snake_case_ : Union[str, Any] = x def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : int = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Any: '''simple docstring''' snake_case_ : str = () snake_case_ : str = () snake_case_ : List[str] = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: snake_case_ : int = all_hidden_states + (hidden_states,) snake_case_ : Any = layer_module( _lowercase , _lowercase , head_mask[i] , _lowercase , _lowercase ) snake_case_ : Dict = layer_outputs[0] if self.output_attentions: snake_case_ : str = all_attentions + (layer_outputs[1],) snake_case_ : Optional[int] = (hidden_states,) if self.output_hidden_states: snake_case_ : Tuple = current_outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : int = current_outputs + (all_attentions,) snake_case_ : Optional[Any] = self.highway[i](_lowercase ) # logits, pooled_output if not self.training: snake_case_ : Tuple = highway_exit[0] snake_case_ : List[str] = entropy(_lowercase ) snake_case_ : str = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy snake_case_ : Union[str, Any] = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: snake_case_ : List[Any] = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(_lowercase , i + 1 ) else: snake_case_ : Dict = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: snake_case_ : Dict = all_hidden_states + (hidden_states,) snake_case_ : str = (hidden_states,) if self.output_hidden_states: snake_case_ : List[Any] = outputs + (all_hidden_states,) if self.output_attentions: snake_case_ : Union[str, Any] = outputs + (all_attentions,) snake_case_ : List[str] = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( '''The Bert Model transformer with early exiting (DeeBERT). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> Union[str, Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config snake_case_ : int = BertEmbeddings(_lowercase ) snake_case_ : Tuple = DeeBertEncoder(_lowercase ) snake_case_ : int = BertPooler(_lowercase ) self.init_weights() def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' self.encoder.init_highway_pooler(self.pooler ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return self.embeddings.word_embeddings def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = value def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(_lowercase ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' if input_ids is not None and inputs_embeds is not None: raise ValueError("""You cannot specify both input_ids and inputs_embeds at the same time""" ) elif input_ids is not None: snake_case_ : Dict = input_ids.size() elif inputs_embeds is not None: snake_case_ : Optional[int] = inputs_embeds.size()[:-1] else: raise ValueError("""You have to specify either input_ids or inputs_embeds""" ) snake_case_ : int = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: snake_case_ : Dict = torch.ones(_lowercase , device=_lowercase ) if encoder_attention_mask is None: snake_case_ : Tuple = torch.ones(_lowercase , device=_lowercase ) if token_type_ids is None: snake_case_ : Any = torch.zeros(_lowercase , dtype=torch.long , device=_lowercase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. snake_case_ : torch.Tensor = self.get_extended_attention_mask(_lowercase , _lowercase , _lowercase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: snake_case_ : List[str] = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: snake_case_ : Any = encoder_attention_mask[:, None, None, :] snake_case_ : List[str] = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility snake_case_ : List[str] = (1.0 - encoder_extended_attention_mask) * -1_0000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] snake_case_ : int = self.get_head_mask(_lowercase , self.config.num_hidden_layers ) snake_case_ : List[str] = self.embeddings( input_ids=_lowercase , position_ids=_lowercase , token_type_ids=_lowercase , inputs_embeds=_lowercase ) snake_case_ : List[str] = self.encoder( _lowercase , attention_mask=_lowercase , head_mask=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , ) snake_case_ : Optional[Any] = encoder_outputs[0] snake_case_ : Union[str, Any] = self.pooler(_lowercase ) snake_case_ : Optional[Any] = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ : Tuple = message snake_case_ : str = exit_layer # start from 1! class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__() snake_case_ : str = BertPooler(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Dict = nn.Linear(config.hidden_size , config.num_labels ) def UpperCAmelCase__ ( self , _lowercase ) -> Dict: '''simple docstring''' snake_case_ : List[str] = encoder_outputs[0] snake_case_ : List[Any] = self.pooler(_lowercase ) # "return" pooler_output # BertModel snake_case_ : Tuple = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification snake_case_ : Union[str, Any] = bmodel_output[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : List[str] = self.classifier(_lowercase ) return logits, pooled_output @add_start_docstrings( '''Bert Model (with early exiting - DeeBERT) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Union[str, Any] = config.num_labels snake_case_ : Tuple = config.num_hidden_layers snake_case_ : Any = DeeBertModel(_lowercase ) snake_case_ : Optional[int] = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : Tuple = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> int: '''simple docstring''' snake_case_ : int = self.num_layers try: snake_case_ : Any = self.bert( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits snake_case_ : str = outputs[1] snake_case_ : Optional[int] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : List[Any] = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : Optional[int] = e.message snake_case_ : Dict = e.exit_layer snake_case_ : Optional[Any] = outputs[0] if not self.training: snake_case_ : int = entropy(_lowercase ) snake_case_ : int = [] snake_case_ : List[str] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : Dict = [] for highway_exit in outputs[-1]: snake_case_ : List[Any] = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : List[Any] = MSELoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Dict = CrossEntropyLoss() snake_case_ : Optional[int] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : List[str] = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : str = (loss,) + outputs if not self.training: snake_case_ : Optional[int] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : str = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
58
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = { '''kssteven/ibert-roberta-base''': '''https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json''', '''kssteven/ibert-roberta-large''': '''https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json''', '''kssteven/ibert-roberta-large-mnli''': ( '''https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json''' ), } class _lowerCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" UpperCAmelCase_ : Optional[int] ="ibert" def __init__( self , UpperCAmelCase=30522 , UpperCAmelCase=768 , UpperCAmelCase=12 , UpperCAmelCase=12 , UpperCAmelCase=3072 , UpperCAmelCase="gelu" , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=512 , UpperCAmelCase=2 , UpperCAmelCase=0.02 , UpperCAmelCase=1E-12 , UpperCAmelCase=1 , UpperCAmelCase=0 , UpperCAmelCase=2 , UpperCAmelCase="absolute" , UpperCAmelCase=False , UpperCAmelCase="none" , **UpperCAmelCase , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) __snake_case : List[Any] = vocab_size __snake_case : str = hidden_size __snake_case : str = num_hidden_layers __snake_case : Dict = num_attention_heads __snake_case : Optional[Any] = hidden_act __snake_case : Any = intermediate_size __snake_case : Any = hidden_dropout_prob __snake_case : List[str] = attention_probs_dropout_prob __snake_case : List[str] = max_position_embeddings __snake_case : List[Any] = type_vocab_size __snake_case : Optional[int] = initializer_range __snake_case : Optional[Any] = layer_norm_eps __snake_case : int = position_embedding_type __snake_case : int = quant_mode __snake_case : List[Any] = force_dequant class _lowerCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": __snake_case : Optional[Any] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __snake_case : Optional[Any] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
243
"""simple docstring""" # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' return 1 / (1 + np.exp(-z )) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int ): '''simple docstring''' return (-y * np.log(__UpperCamelCase ) - (1 - y) * np.log(1 - h )).mean() def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : List[Any] , __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Optional[int] = np.dot(__UpperCamelCase , __UpperCamelCase ) return np.sum(y * scores - np.log(1 + np.exp(__UpperCamelCase ) ) ) def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : int=7_0_0_0_0 ): '''simple docstring''' snake_case_ : Dict = np.zeros(x.shape[1] ) for iterations in range(__UpperCamelCase ): snake_case_ : Any = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Optional[Any] = np.dot(x.T , h - y ) / y.size snake_case_ : str = theta - alpha * gradient # updating the weights snake_case_ : int = np.dot(__UpperCamelCase , __UpperCamelCase ) snake_case_ : List[str] = sigmoid_function(__UpperCamelCase ) snake_case_ : Dict = cost_function(__UpperCamelCase , __UpperCamelCase ) if iterations % 1_0_0 == 0: print(F'loss: {j} \t' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": __lowerCAmelCase : Any = datasets.load_iris() __lowerCAmelCase : List[Any] = iris.data[:, :2] __lowerCAmelCase : Tuple = (iris.target != 0) * 1 __lowerCAmelCase : Any = 0.1 __lowerCAmelCase : List[Any] = logistic_reg(alpha, x, y, max_iterations=7_0000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' return sigmoid_function( np.dot(__UpperCamelCase , __UpperCamelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((__lowerCAmelCase) , (__lowerCAmelCase)) : Union[str, Any] = (x[:, 0].min(), x[:, 0].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Tuple = (x[:, 1].min(), x[:, 1].max()) ((__lowerCAmelCase) , (__lowerCAmelCase)) : Optional[Any] = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) __lowerCAmelCase : Any = np.c_[xxa.ravel(), xxa.ravel()] __lowerCAmelCase : Optional[int] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
58
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : List[Any] , **_snake_case : Union[str, Any] ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Tuple , **_snake_case : Tuple ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : List[Any] , **_snake_case : Optional[int] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : List[str] , **_snake_case : str ) -> List[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Optional[int] , **_snake_case : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Any , **_snake_case : Optional[int] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[Any] , *_snake_case : Dict , **_snake_case : Optional[Any] ) -> int: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : Union[str, Any] , **_snake_case : Dict ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : List[str] , **_snake_case : Tuple ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Dict , *_snake_case : int , **_snake_case : Optional[Any] ) -> List[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Dict , **_snake_case : Dict ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : List[str] , **_snake_case : Optional[int] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : int , **_snake_case : Optional[int] ) -> List[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Dict , **_snake_case : Optional[int] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : Optional[int] , **_snake_case : List[str] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : int , *_snake_case : Any , **_snake_case : Optional[int] ) -> Optional[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : Optional[Any] , **_snake_case : Dict ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : int , **_snake_case : Any ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Any , *_snake_case : Dict , **_snake_case : Dict ) -> int: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : Optional[Any] , **_snake_case : Optional[int] ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Tuple , **_snake_case : int ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Any , *_snake_case : Tuple , **_snake_case : str ) -> Optional[int]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : int , **_snake_case : List[str] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : Optional[Any] , **_snake_case : Dict ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : List[str] , **_snake_case : Optional[int] ) -> List[str]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : str , **_snake_case : Tuple ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : Dict , **_snake_case : Union[str, Any] ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Dict , *_snake_case : Dict , **_snake_case : List[Any] ) -> int: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : int , **_snake_case : Union[str, Any] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : int , **_snake_case : Optional[int] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : Union[str, Any] , **_snake_case : Dict ) -> Optional[int]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , *_snake_case : Dict , **_snake_case : int ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : Tuple , **_snake_case : Tuple ) -> Tuple: """simple docstring""" requires_backends(cls , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) def A_ (*__a , **__a ): '''simple docstring''' requires_backends(__UpperCamelCase , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : int , **_snake_case : Any ) -> Any: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : str , **_snake_case : Optional[Any] ) -> List[str]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : Tuple , **_snake_case : int ) -> Any: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : List[Any] , **_snake_case : Union[str, Any] ) -> int: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : List[str] , **_snake_case : int ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Tuple , **_snake_case : str ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Dict , *_snake_case : Dict , **_snake_case : Optional[int] ) -> str: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : List[Any] , **_snake_case : Union[str, Any] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Optional[Any] , **_snake_case : Optional[int] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Union[str, Any] , *_snake_case : Optional[Any] , **_snake_case : Any ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : int , **_snake_case : Union[str, Any] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : Any , **_snake_case : Union[str, Any] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : int , **_snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : int , **_snake_case : Tuple ) -> List[str]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : List[Any] , **_snake_case : int ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : str , **_snake_case : Union[str, Any] ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : int , **_snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : str , **_snake_case : List[str] ) -> List[str]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Union[str, Any] , *_snake_case : Tuple , **_snake_case : str ) -> List[str]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : List[Any] , **_snake_case : int ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : int , **_snake_case : Dict ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : int , *_snake_case : Any , **_snake_case : str ) -> List[str]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : List[str] , **_snake_case : Optional[int] ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : List[str] , **_snake_case : Tuple ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[Any] , *_snake_case : List[Any] , **_snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : List[Any] , **_snake_case : Dict ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : Any , **_snake_case : Union[str, Any] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : Optional[Any] , **_snake_case : Optional[Any] ) -> List[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : List[Any] , **_snake_case : Optional[int] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : int , **_snake_case : Any ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : Dict , **_snake_case : List[str] ) -> str: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Optional[int] , **_snake_case : Any ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : Optional[Any] , **_snake_case : Tuple ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Tuple , *_snake_case : List[str] , **_snake_case : Dict ) -> Dict: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : str , **_snake_case : str ) -> Any: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : Dict , **_snake_case : Optional[Any] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : Dict , **_snake_case : List[str] ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : int , **_snake_case : str ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , *_snake_case : str , **_snake_case : int ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Any , *_snake_case : Optional[Any] , **_snake_case : Union[str, Any] ) -> int: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Optional[Any] , **_snake_case : List[str] ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : str , **_snake_case : int ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : List[str] , **_snake_case : Union[str, Any] ) -> Any: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , *_snake_case : List[str] , **_snake_case : int ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : Dict , **_snake_case : List[Any] ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[Any] , *_snake_case : Optional[Any] , **_snake_case : Optional[Any] ) -> Optional[int]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : List[str] , **_snake_case : Dict ) -> Any: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : List[Any] , **_snake_case : Optional[Any] ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Tuple , *_snake_case : Optional[Any] , **_snake_case : Any ) -> Any: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Union[str, Any] , **_snake_case : List[Any] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : Tuple , **_snake_case : Any ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : Any , **_snake_case : int ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : List[Any] , **_snake_case : int ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : Any , **_snake_case : Tuple ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : Dict , **_snake_case : Any ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : List[Any] , **_snake_case : Optional[Any] ) -> List[str]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : Optional[Any] , **_snake_case : List[str] ) -> Any: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : Optional[Any] , **_snake_case : Optional[Any] ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , *_snake_case : Optional[int] , **_snake_case : str ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : Dict , **_snake_case : Tuple ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Tuple , *_snake_case : Tuple , **_snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : int , **_snake_case : List[str] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Tuple , *_snake_case : Union[str, Any] , **_snake_case : Optional[Any] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[Any] , *_snake_case : Dict , **_snake_case : Union[str, Any] ) -> Any: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : int , **_snake_case : Tuple ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Optional[int] , **_snake_case : List[Any] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Any , *_snake_case : str , **_snake_case : str ) -> Union[str, Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Union[str, Any] , **_snake_case : Any ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : List[Any] , **_snake_case : Dict ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : str , **_snake_case : Any ) -> List[str]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , *_snake_case : Union[str, Any] , **_snake_case : str ) -> Tuple: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : int , **_snake_case : Optional[Any] ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[Any] , *_snake_case : List[Any] , **_snake_case : str ) -> Union[str, Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : Dict , **_snake_case : Any ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : Any , **_snake_case : Optional[Any] ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Any , *_snake_case : Optional[int] , **_snake_case : Any ) -> Dict: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : List[Any] , **_snake_case : Optional[Any] ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Any , **_snake_case : List[Any] ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : Any , **_snake_case : int ) -> Optional[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Optional[int] , **_snake_case : Any ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : Union[str, Any] , **_snake_case : Optional[Any] ) -> Any: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : Optional[Any] , **_snake_case : str ) -> Optional[int]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Tuple , **_snake_case : List[str] ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : Dict , **_snake_case : Any ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[str] , *_snake_case : Dict , **_snake_case : Optional[int] ) -> Tuple: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : Any , **_snake_case : List[Any] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Dict , **_snake_case : Any ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[Any] , *_snake_case : Optional[int] , **_snake_case : Any ) -> List[Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Union[str, Any] , *_snake_case : List[str] , **_snake_case : List[Any] ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : int , **_snake_case : Optional[int] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : List[Any] , **_snake_case : Tuple ) -> Dict: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : List[Any] , **_snake_case : str ) -> Any: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : Dict , **_snake_case : int ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Optional[int] , *_snake_case : Union[str, Any] , **_snake_case : Dict ) -> str: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Any , **_snake_case : Union[str, Any] ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : int , **_snake_case : int ) -> Tuple: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Union[str, Any] , *_snake_case : Optional[Any] , **_snake_case : List[Any] ) -> str: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : str , *_snake_case : Optional[int] , **_snake_case : Tuple ) -> str: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Dict , *_snake_case : Union[str, Any] , **_snake_case : Any ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : List[Any] , *_snake_case : str , **_snake_case : Dict ) -> Dict: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : str , **_snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[str] , *_snake_case : Dict , **_snake_case : Optional[int] ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Any , *_snake_case : Dict , **_snake_case : Union[str, Any] ) -> List[str]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : Union[str, Any] , **_snake_case : Tuple ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : Any , **_snake_case : Dict ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : str , *_snake_case : str , **_snake_case : int ) -> Union[str, Any]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : List[Any] , **_snake_case : Dict ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[int] , *_snake_case : List[str] , **_snake_case : Any ) -> int: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Union[str, Any] , *_snake_case : List[str] , **_snake_case : str ) -> Optional[int]: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : List[Any] , **_snake_case : Union[str, Any] ) -> List[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Dict , **_snake_case : Optional[Any] ) -> List[str]: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : Dict , *_snake_case : Dict , **_snake_case : List[Any] ) -> int: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : List[Any] , *_snake_case : str , **_snake_case : Optional[Any] ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Optional[Any] , *_snake_case : Union[str, Any] , **_snake_case : Union[str, Any] ) -> Dict: """simple docstring""" requires_backends(cls , ["torch"] ) class __lowerCAmelCase ( metaclass=SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : int , *_snake_case : Dict , **_snake_case : Any ) -> str: """simple docstring""" requires_backends(self , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : int , *_snake_case : Optional[int] , **_snake_case : int ) -> Optional[Any]: """simple docstring""" requires_backends(cls , ["torch"] ) @classmethod def lowerCamelCase__ ( cls : Any , *_snake_case : Optional[int] , **_snake_case : Dict ) -> Tuple: """simple docstring""" requires_backends(cls , ["torch"] )
115
"""simple docstring""" import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput __lowerCAmelCase : Tuple = '''scheduler_config.json''' class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 1 _lowerCamelCase = 2 _lowerCamelCase = 3 _lowerCamelCase = 4 _lowerCamelCase = 5 @dataclass class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = 42 class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = SCHEDULER_CONFIG_NAME _lowerCamelCase = ['''dtype'''] _lowerCamelCase = [] _lowerCamelCase = True @classmethod def UpperCAmelCase__ ( cls , _lowercase = None , _lowercase = None , _lowercase=False , **_lowercase , ) -> Any: '''simple docstring''' snake_case_ , snake_case_ : int = cls.load_config( pretrained_model_name_or_path=_lowercase , subfolder=_lowercase , return_unused_kwargs=_lowercase , **_lowercase , ) snake_case_ , snake_case_ : Dict = cls.from_config(_lowercase , return_unused_kwargs=_lowercase , **_lowercase ) if hasattr(_lowercase , """create_state""" ) and getattr(_lowercase , """has_state""" , _lowercase ): snake_case_ : Any = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def UpperCAmelCase__ ( self , _lowercase , _lowercase = False , **_lowercase ) -> Optional[Any]: '''simple docstring''' self.save_config(save_directory=_lowercase , push_to_hub=_lowercase , **_lowercase ) @property def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' return self._get_compatibles() @classmethod def UpperCAmelCase__ ( cls ) -> Dict: '''simple docstring''' snake_case_ : Union[str, Any] = list(set([cls.__name__] + cls._compatibles ) ) snake_case_ : str = importlib.import_module(__name__.split(""".""" )[0] ) snake_case_ : Optional[int] = [ getattr(_lowercase , _lowercase ) for c in compatible_classes_str if hasattr(_lowercase , _lowercase ) ] return compatible_classes def __lowerCAmelCase ( __UpperCamelCase : jnp.ndarray , __UpperCamelCase : Tuple[int] ): '''simple docstring''' assert len(__UpperCamelCase ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(__UpperCamelCase ) - x.ndim) ) , __UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Any=0.999 , __UpperCamelCase : Optional[int]=jnp.floataa ): '''simple docstring''' def alpha_bar(__UpperCamelCase : Optional[int] ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 snake_case_ : Optional[Any] = [] for i in range(__UpperCamelCase ): snake_case_ : Dict = i / num_diffusion_timesteps snake_case_ : Union[str, Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(__UpperCamelCase ) / alpha_bar(__UpperCamelCase ) , __UpperCamelCase ) ) return jnp.array(__UpperCamelCase , dtype=__UpperCamelCase ) @flax.struct.dataclass class _lowerCAmelCase : """simple docstring""" _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 @classmethod def UpperCAmelCase__ ( cls , _lowercase ) -> int: '''simple docstring''' snake_case_ : Any = scheduler.config if config.trained_betas is not None: snake_case_ : Optional[Any] = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": snake_case_ : int = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case_ : str = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case_ : int = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( f'beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}' ) snake_case_ : Optional[Any] = 1.0 - betas snake_case_ : Any = jnp.cumprod(_lowercase , axis=0 ) return cls( alphas=_lowercase , betas=_lowercase , alphas_cumprod=_lowercase , ) def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ : Tuple = state.alphas_cumprod snake_case_ : Optional[int] = alphas_cumprod[timesteps] ** 0.5 snake_case_ : Dict = sqrt_alpha_prod.flatten() snake_case_ : int = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) snake_case_ : Optional[Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 snake_case_ : Dict = sqrt_one_minus_alpha_prod.flatten() snake_case_ : Tuple = broadcast_to_shape_from_left(__UpperCamelCase , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : str = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def __lowerCAmelCase ( __UpperCamelCase : CommonSchedulerState , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray , __UpperCamelCase : jnp.ndarray ): '''simple docstring''' snake_case_ , snake_case_ : List[Any] = get_sqrt_alpha_prod(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ : Any = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
58
0
import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError("""At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training""") # TF training parameters lowercase_ = False lowercase_ = False def a__ ( snake_case ): """simple docstring""" return TrainCommand(__UpperCamelCase ) class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @staticmethod def UpperCAmelCase__ ( _A : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = parser.add_parser('''train''' , help='''CLI tool to train a model on a task.''' ) train_parser.add_argument( '''--train_data''' , type=_lowercase , required=_lowercase , help='''path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.''' , ) train_parser.add_argument( '''--column_label''' , type=_lowercase , default=0 , help='''Column of the dataset csv file with example labels.''' ) train_parser.add_argument( '''--column_text''' , type=_lowercase , default=1 , help='''Column of the dataset csv file with example texts.''' ) train_parser.add_argument( '''--column_id''' , type=_lowercase , default=2 , help='''Column of the dataset csv file with example ids.''' ) train_parser.add_argument( '''--skip_first_row''' , action='''store_true''' , help='''Skip the first row of the csv file (headers).''' ) train_parser.add_argument('''--validation_data''' , type=_lowercase , default='''''' , help='''path to validation dataset.''' ) train_parser.add_argument( '''--validation_split''' , type=_lowercase , default=0.1 , help='''if validation dataset is not provided, fraction of train dataset to use as validation dataset.''' , ) train_parser.add_argument('''--output''' , type=_lowercase , default='''./''' , help='''path to saved the trained model.''' ) train_parser.add_argument( '''--task''' , type=_lowercase , default='''text_classification''' , help='''Task to train the model on.''' ) train_parser.add_argument( '''--model''' , type=_lowercase , default='''bert-base-uncased''' , help='''Model\'s name or path to stored model.''' ) train_parser.add_argument('''--train_batch_size''' , type=_lowercase , default=32 , help='''Batch size for training.''' ) train_parser.add_argument('''--valid_batch_size''' , type=_lowercase , default=64 , help='''Batch size for validation.''' ) train_parser.add_argument('''--learning_rate''' , type=_lowercase , default=3e-5 , help='''Learning rate.''' ) train_parser.add_argument('''--adam_epsilon''' , type=_lowercase , default=1e-08 , help='''Epsilon for Adam optimizer.''' ) train_parser.set_defaults(func=_lowercase ) def __init__( self : List[Any] , _A : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger('''transformers-cli/training''' ) __SCREAMING_SNAKE_CASE : Dict = """tf""" if is_tf_available() else """torch""" os.makedirs(args.output , exist_ok=_lowercase ) __SCREAMING_SNAKE_CASE : List[Any] = args.output __SCREAMING_SNAKE_CASE : Optional[int] = args.column_label __SCREAMING_SNAKE_CASE : List[str] = args.column_text __SCREAMING_SNAKE_CASE : int = args.column_id self.logger.info(F'''Loading {args.task} pipeline for {args.model}''' ) if args.task == "text_classification": __SCREAMING_SNAKE_CASE : Any = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(F'''Loading dataset from {args.train_data}''' ) __SCREAMING_SNAKE_CASE : Tuple = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) __SCREAMING_SNAKE_CASE : Tuple = None if args.validation_data: self.logger.info(F'''Loading validation dataset from {args.validation_data}''' ) __SCREAMING_SNAKE_CASE : List[Any] = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) __SCREAMING_SNAKE_CASE : Dict = args.validation_split __SCREAMING_SNAKE_CASE : Tuple = args.train_batch_size __SCREAMING_SNAKE_CASE : List[Any] = args.valid_batch_size __SCREAMING_SNAKE_CASE : List[Any] = args.learning_rate __SCREAMING_SNAKE_CASE : str = args.adam_epsilon def UpperCAmelCase__ ( self : List[Any] ): """simple docstring""" if self.framework == "tf": return self.run_tf() return self.run_torch() def UpperCAmelCase__ ( self : Dict ): """simple docstring""" raise NotImplementedError def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
74
"""simple docstring""" from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( '''The RoBERTa Model transformer with early exiting (DeeRoBERTa). ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> Optional[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : str = RobertaEmbeddings(_lowercase ) self.init_weights() @add_start_docstrings( '''RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top, also takes care of multi-layer training. ''' , SCREAMING_SNAKE_CASE__ , ) class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = RobertaConfig _lowerCamelCase = '''roberta''' def __init__( self , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__(_lowercase ) snake_case_ : Optional[Any] = config.num_labels snake_case_ : Dict = config.num_hidden_layers snake_case_ : str = DeeRobertaModel(_lowercase ) snake_case_ : Dict = nn.Dropout(config.hidden_dropout_prob ) snake_case_ : List[str] = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(_lowercase ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=-1 , _lowercase=False , ) -> Tuple: '''simple docstring''' snake_case_ : Any = self.num_layers try: snake_case_ : int = self.roberta( _lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , ) snake_case_ : str = outputs[1] snake_case_ : Union[str, Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.classifier(_lowercase ) snake_case_ : Dict = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: snake_case_ : List[Any] = e.message snake_case_ : Union[str, Any] = e.exit_layer snake_case_ : Dict = outputs[0] if not self.training: snake_case_ : Dict = entropy(_lowercase ) snake_case_ : Optional[int] = [] snake_case_ : Union[str, Any] = [] if labels is not None: if self.num_labels == 1: # We are doing regression snake_case_ : Dict = MSELoss() snake_case_ : Dict = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Union[str, Any] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits snake_case_ : int = [] for highway_exit in outputs[-1]: snake_case_ : Tuple = highway_exit[0] if not self.training: highway_logits_all.append(_lowercase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression snake_case_ : Optional[int] = MSELoss() snake_case_ : Optional[Any] = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: snake_case_ : Optional[int] = CrossEntropyLoss() snake_case_ : Union[str, Any] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_lowercase ) if train_highway: snake_case_ : Dict = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: snake_case_ : List[str] = (loss,) + outputs if not self.training: snake_case_ : Optional[Any] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: snake_case_ : Tuple = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
58
0
'''simple docstring''' import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def a_ ( __snake_case : bytes , __snake_case : int ) -> Dict: """simple docstring""" lowerCamelCase_ =F'''{sampling_rate}''' lowerCamelCase_ ="""1""" lowerCamelCase_ ="""f32le""" lowerCamelCase_ =[ """ffmpeg""", """-i""", """pipe:0""", """-ac""", ac, """-ar""", ar, """-f""", format_for_conversion, """-hide_banner""", """-loglevel""", """quiet""", """pipe:1""", ] try: with subprocess.Popen(__UpperCamelCase , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: lowerCamelCase_ =ffmpeg_process.communicate(__UpperCamelCase ) except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to load audio files from filename''' ) from error lowerCamelCase_ =output_stream[0] lowerCamelCase_ =np.frombuffer(__UpperCamelCase , np.floataa ) if audio.shape[0] == 0: raise ValueError('''Malformed soundfile''' ) return audio def a_ ( __snake_case : int , __snake_case : float , __snake_case : str = "f32le" , ) -> Optional[int]: """simple docstring""" lowerCamelCase_ =F'''{sampling_rate}''' lowerCamelCase_ ="""1""" if format_for_conversion == "s16le": lowerCamelCase_ =2 elif format_for_conversion == "f32le": lowerCamelCase_ =4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) lowerCamelCase_ =platform.system() if system == "Linux": lowerCamelCase_ ="""alsa""" lowerCamelCase_ ="""default""" elif system == "Darwin": lowerCamelCase_ ="""avfoundation""" lowerCamelCase_ =""":0""" elif system == "Windows": lowerCamelCase_ ="""dshow""" lowerCamelCase_ ="""default""" lowerCamelCase_ =[ """ffmpeg""", """-f""", format_, """-i""", input_, """-ac""", ac, """-ar""", ar, """-f""", format_for_conversion, """-fflags""", """nobuffer""", """-hide_banner""", """-loglevel""", """quiet""", """pipe:1""", ] lowerCamelCase_ =int(round(sampling_rate * chunk_length_s ) ) * size_of_sample lowerCamelCase_ =_ffmpeg_stream(__UpperCamelCase , __UpperCamelCase ) for item in iterator: yield item def a_ ( __snake_case : int , __snake_case : float , __snake_case : Optional[int] = None , __snake_case : Optional[Union[Tuple[float, float], float]] = None , __snake_case : str = "f32le" , ) -> List[Any]: """simple docstring""" if stream_chunk_s is not None: lowerCamelCase_ =stream_chunk_s else: lowerCamelCase_ =chunk_length_s lowerCamelCase_ =ffmpeg_microphone(__UpperCamelCase , __UpperCamelCase , format_for_conversion=__UpperCamelCase ) if format_for_conversion == "s16le": lowerCamelCase_ =np.intaa lowerCamelCase_ =2 elif format_for_conversion == "f32le": lowerCamelCase_ =np.floataa lowerCamelCase_ =4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: lowerCamelCase_ =chunk_length_s / 6 lowerCamelCase_ =int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__UpperCamelCase , (int, float) ): lowerCamelCase_ =[stride_length_s, stride_length_s] lowerCamelCase_ =int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample lowerCamelCase_ =int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample lowerCamelCase_ =datetime.datetime.now() lowerCamelCase_ =datetime.timedelta(seconds=__UpperCamelCase ) for item in chunk_bytes_iter(__UpperCamelCase , __UpperCamelCase , stride=(stride_left, stride_right) , stream=__UpperCamelCase ): # Put everything back in numpy scale lowerCamelCase_ =np.frombuffer(item['''raw'''] , dtype=__UpperCamelCase ) lowerCamelCase_ =( item["""stride"""][0] // size_of_sample, item["""stride"""][1] // size_of_sample, ) lowerCamelCase_ =sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def a_ ( __snake_case : Optional[int] , __snake_case : int , __snake_case : Tuple[int, int] , __snake_case : bool = False ) -> Dict: """simple docstring""" lowerCamelCase_ =B"""""" lowerCamelCase_ =stride if stride_left + stride_right >= chunk_len: raise ValueError( F'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) lowerCamelCase_ =0 for raw in iterator: acc += raw if stream and len(__UpperCamelCase ) < chunk_len: lowerCamelCase_ =(_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__UpperCamelCase ) >= chunk_len: # We are flushing the accumulator lowerCamelCase_ =(_stride_left, stride_right) lowerCamelCase_ ={"""raw""": acc[:chunk_len], """stride""": stride} if stream: lowerCamelCase_ =False yield item lowerCamelCase_ =stride_left lowerCamelCase_ =acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__UpperCamelCase ) > stride_left: lowerCamelCase_ ={"""raw""": acc, """stride""": (_stride_left, 0)} if stream: lowerCamelCase_ =False yield item def a_ ( __snake_case : Any , __snake_case : int ) -> Optional[int]: """simple docstring""" lowerCamelCase_ =2**24 # 16Mo try: with subprocess.Popen(__UpperCamelCase , stdout=subprocess.PIPE , bufsize=__UpperCamelCase ) as ffmpeg_process: while True: lowerCamelCase_ =ffmpeg_process.stdout.read(__UpperCamelCase ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to stream audio files from filename''' ) from error
676
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int , __UpperCamelCase : int , __UpperCamelCase : list[int] ): '''simple docstring''' if graph[path[curr_ind - 1]][next_ver] == 0: return False # 2. Validate that next vertex is not already in path return not any(vertex == next_ver for vertex in path ) def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : list[int] , __UpperCamelCase : int ): '''simple docstring''' if curr_ind == len(__UpperCamelCase ): # return whether path exists between current and starting vertices return graph[path[curr_ind - 1]][path[0]] == 1 # Recursive Step for next_ver in range(0 , len(__UpperCamelCase ) ): if valid_connection(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): # Insert current vertex into path as next transition snake_case_ : List[str] = next_ver # Validate created path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , curr_ind + 1 ): return True # Backtrack snake_case_ : Tuple = -1 return False def __lowerCAmelCase ( __UpperCamelCase : list[list[int]] , __UpperCamelCase : int = 0 ): '''simple docstring''' snake_case_ : Tuple = [-1] * (len(__UpperCamelCase ) + 1) # initialize start and end of path with starting index snake_case_ : Optional[int] = start_index # evaluate and if we find answer return path either return empty array return path if util_hamilton_cycle(__UpperCamelCase , __UpperCamelCase , 1 ) else []
58
0
from typing import Union import fire import torch from tqdm import tqdm def _A ( SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE = "cpu" ,SCREAMING_SNAKE_CASE = None ): UpperCAmelCase__: str = torch.load(__UpperCamelCase ,map_location=__UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(__UpperCamelCase ,torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) UpperCAmelCase__: Tuple = v.half() if save_path is None: # overwrite src_path UpperCAmelCase__: List[str] = src_path torch.save(__UpperCamelCase ,__UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
113
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = ['''image_processor''', '''tokenizer'''] _lowerCamelCase = '''BlipImageProcessor''' _lowerCamelCase = '''AutoTokenizer''' def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__(_lowercase , _lowercase ) # add QFormer tokenizer snake_case_ : List[str] = qformer_tokenizer def __call__( self , _lowercase = None , _lowercase = None , _lowercase = True , _lowercase = False , _lowercase = None , _lowercase = None , _lowercase = 0 , _lowercase = None , _lowercase = None , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = False , _lowercase = True , _lowercase = None , **_lowercase , ) -> BatchFeature: '''simple docstring''' if images is None and text is None: raise ValueError("""You have to specify at least images or text.""" ) snake_case_ : Optional[Any] = BatchFeature() if text is not None: snake_case_ : List[str] = self.tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) encoding.update(_lowercase ) snake_case_ : Union[str, Any] = self.qformer_tokenizer( text=_lowercase , add_special_tokens=_lowercase , padding=_lowercase , truncation=_lowercase , max_length=_lowercase , stride=_lowercase , pad_to_multiple_of=_lowercase , return_attention_mask=_lowercase , return_overflowing_tokens=_lowercase , return_special_tokens_mask=_lowercase , return_offsets_mapping=_lowercase , return_token_type_ids=_lowercase , return_length=_lowercase , verbose=_lowercase , return_tensors=_lowercase , **_lowercase , ) snake_case_ : List[str] = qformer_text_encoding.pop("""input_ids""" ) snake_case_ : Union[str, Any] = qformer_text_encoding.pop("""attention_mask""" ) if images is not None: snake_case_ : Tuple = self.image_processor(_lowercase , return_tensors=_lowercase ) encoding.update(_lowercase ) return encoding def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> List[Any]: '''simple docstring''' return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def UpperCAmelCase__ ( self , *_lowercase , **_lowercase ) -> Dict: '''simple docstring''' return self.tokenizer.decode(*_lowercase , **_lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' snake_case_ : str = self.tokenizer.model_input_names snake_case_ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def UpperCAmelCase__ ( self , _lowercase , **_lowercase ) -> Optional[int]: '''simple docstring''' if os.path.isfile(_lowercase ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(_lowercase , exist_ok=_lowercase ) snake_case_ : int = os.path.join(_lowercase , """qformer_tokenizer""" ) self.qformer_tokenizer.save_pretrained(_lowercase ) return super().save_pretrained(_lowercase , **_lowercase ) @classmethod def UpperCAmelCase__ ( cls , _lowercase , **_lowercase ) -> int: '''simple docstring''' snake_case_ : List[str] = AutoTokenizer.from_pretrained(_lowercase , subfolder="""qformer_tokenizer""" ) snake_case_ : Union[str, Any] = cls._get_arguments_from_pretrained(_lowercase , **_lowercase ) args.append(_lowercase ) return cls(*_lowercase )
58
0
"""simple docstring""" import argparse import os import re __UpperCamelCase : Any = '''src/diffusers''' # Pattern that looks at the indentation in a line. __UpperCamelCase : Any = re.compile(R'''^(\s*)\S''') # Pattern that matches `"key":" and puts `key` in group 0. __UpperCamelCase : Union[str, Any] = re.compile(R'''^\s*"([^"]+)":''') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. __UpperCamelCase : Tuple = re.compile(R'''^\s*_import_structure\["([^"]+)"\]''') # Pattern that matches `"key",` and puts `key` in group 0. __UpperCamelCase : Any = re.compile(R'''^\s*"([^"]+)",\s*$''') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. __UpperCamelCase : List[Any] = re.compile(R'''\[([^\]]+)\]''') def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Any = _re_indent.search(__UpperCamelCase ) return "" if search is None else search.groups()[0] def __SCREAMING_SNAKE_CASE ( A_ , A_="" , A_=None , A_=None ): lowerCAmelCase__ : int = 0 lowerCAmelCase__ : Dict = code.split('''\n''' ) if start_prompt is not None: while not lines[index].startswith(__UpperCamelCase ): index += 1 lowerCAmelCase__ : Optional[int] = ["""\n""".join(lines[:index] )] else: lowerCAmelCase__ : List[str] = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). lowerCAmelCase__ : Any = [lines[index]] index += 1 while index < len(__UpperCamelCase ) and (end_prompt is None or not lines[index].startswith(__UpperCamelCase )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(__UpperCamelCase ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ''' ''' ): current_block.append(lines[index] ) blocks.append('''\n'''.join(__UpperCamelCase ) ) if index < len(__UpperCamelCase ) - 1: lowerCAmelCase__ : Union[str, Any] = [lines[index + 1]] index += 1 else: lowerCAmelCase__ : Optional[int] = [] else: blocks.append('''\n'''.join(__UpperCamelCase ) ) lowerCAmelCase__ : Union[str, Any] = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(__UpperCamelCase ) > 0: blocks.append('''\n'''.join(__UpperCamelCase ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(__UpperCamelCase ): blocks.append('''\n'''.join(lines[index:] ) ) return blocks def __SCREAMING_SNAKE_CASE ( A_ ): def _inner(A_ ): return key(__UpperCamelCase ).lower().replace('''_''' , '''''' ) return _inner def __SCREAMING_SNAKE_CASE ( A_ , A_=None ): def noop(A_ ): return x if key is None: lowerCAmelCase__ : Tuple = noop # Constants are all uppercase, they go first. lowerCAmelCase__ : Dict = [obj for obj in objects if key(__UpperCamelCase ).isupper()] # Classes are not all uppercase but start with a capital, they go second. lowerCAmelCase__ : str = [obj for obj in objects if key(__UpperCamelCase )[0].isupper() and not key(__UpperCamelCase ).isupper()] # Functions begin with a lowercase, they go last. lowerCAmelCase__ : int = [obj for obj in objects if not key(__UpperCamelCase )[0].isupper()] lowerCAmelCase__ : Optional[int] = ignore_underscore(__UpperCamelCase ) return sorted(__UpperCamelCase , key=__UpperCamelCase ) + sorted(__UpperCamelCase , key=__UpperCamelCase ) + sorted(__UpperCamelCase , key=__UpperCamelCase ) def __SCREAMING_SNAKE_CASE ( A_ ): def _replace(A_ ): lowerCAmelCase__ : List[Any] = match.groups()[0] if "," not in imports: return f'[{imports}]' lowerCAmelCase__ : Tuple = [part.strip().replace('''\"''' , '''''' ) for part in imports.split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: lowerCAmelCase__ : Optional[Any] = keys[:-1] return "[" + ", ".join([f'"{k}"' for k in sort_objects(__UpperCamelCase )] ) + "]" lowerCAmelCase__ : Optional[Any] = import_statement.split('''\n''' ) if len(__UpperCamelCase ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. lowerCAmelCase__ : Any = 2 if lines[1].strip() == """[""" else 1 lowerCAmelCase__ : Any = [(i, _re_strip_line.search(__UpperCamelCase ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] lowerCAmelCase__ : List[str] = sort_objects(__UpperCamelCase , key=lambda A_ : x[1] ) lowerCAmelCase__ : Union[str, Any] = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(__UpperCamelCase ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: lowerCAmelCase__ : Any = _re_bracket_content.sub(_replace , lines[1] ) else: lowerCAmelCase__ : List[str] = [part.strip().replace('''\"''' , '''''' ) for part in lines[1].split(''',''' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: lowerCAmelCase__ : Tuple = keys[:-1] lowerCAmelCase__ : List[str] = get_indent(lines[1] ) + """, """.join([f'"{k}"' for k in sort_objects(__UpperCamelCase )] ) return "\n".join(__UpperCamelCase ) else: # Finally we have to deal with imports fitting on one line lowerCAmelCase__ : Any = _re_bracket_content.sub(_replace , __UpperCamelCase ) return import_statement def __SCREAMING_SNAKE_CASE ( A_ , A_=True ): with open(__UpperCamelCase , '''r''' ) as f: lowerCAmelCase__ : str = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 lowerCAmelCase__ : int = split_code_in_indented_blocks( __UpperCamelCase , start_prompt='''_import_structure = {''' , end_prompt='''if TYPE_CHECKING:''' ) # We ignore block 0 (everything until start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(__UpperCamelCase ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. lowerCAmelCase__ : List[str] = main_blocks[block_idx] lowerCAmelCase__ : List[str] = block.split('''\n''' ) # Get to the start of the imports. lowerCAmelCase__ : int = 0 while line_idx < len(__UpperCamelCase ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: lowerCAmelCase__ : int = len(__UpperCamelCase ) else: line_idx += 1 if line_idx >= len(__UpperCamelCase ): continue # Ignore beginning and last line: they don't contain anything. lowerCAmelCase__ : Dict = """\n""".join(block_lines[line_idx:-1] ) lowerCAmelCase__ : Tuple = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. lowerCAmelCase__ : List[str] = split_code_in_indented_blocks(__UpperCamelCase , indent_level=__UpperCamelCase ) # We have two categories of import key: list or _import_structure[key].append/extend lowerCAmelCase__ : Dict = _re_direct_key if """_import_structure""" in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. lowerCAmelCase__ : str = [(pattern.search(__UpperCamelCase ).groups()[0] if pattern.search(__UpperCamelCase ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. lowerCAmelCase__ : str = [(i, key) for i, key in enumerate(__UpperCamelCase ) if key is not None] lowerCAmelCase__ : int = [x[0] for x in sorted(__UpperCamelCase , key=lambda A_ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. lowerCAmelCase__ : int = 0 lowerCAmelCase__ : int = [] for i in range(len(__UpperCamelCase ) ): if keys[i] is None: reordered_blocks.append(internal_blocks[i] ) else: lowerCAmelCase__ : str = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reordered_blocks.append(__UpperCamelCase ) count += 1 # And we put our main block back together with its first and last line. lowerCAmelCase__ : Any = """\n""".join(block_lines[:line_idx] + reordered_blocks + [block_lines[-1]] ) if code != "\n".join(__UpperCamelCase ): if check_only: return True else: print(f'Overwriting {file}.' ) with open(__UpperCamelCase , '''w''' ) as f: f.write('''\n'''.join(__UpperCamelCase ) ) def __SCREAMING_SNAKE_CASE ( A_=True ): lowerCAmelCase__ : Any = [] for root, _, files in os.walk(__UpperCamelCase ): if "__init__.py" in files: lowerCAmelCase__ : Tuple = sort_imports(os.path.join(__UpperCamelCase , '''__init__.py''' ) , check_only=__UpperCamelCase ) if result: lowerCAmelCase__ : List[str] = [os.path.join(__UpperCamelCase , '''__init__.py''' )] if len(__UpperCamelCase ) > 0: raise ValueError(f'Would overwrite {len(__UpperCamelCase )} files, run `make style`.' ) if __name__ == "__main__": __UpperCamelCase : Tuple = argparse.ArgumentParser() parser.add_argument('''--check_only''', action='''store_true''', help='''Whether to only check or fix style.''') __UpperCamelCase : Dict = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
450
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCAmelCase : List[Any] = { '''configuration_vivit''': ['''VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VivitConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = ['''VivitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = [ '''VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VivitModel''', '''VivitPreTrainedModel''', '''VivitForVideoClassification''', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
58
0
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class a_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = False , lowercase_ = None , lowercase_ = None , **lowercase_ , ) -> Optional[Any]: '''simple docstring''' super().__init__( features=_lowercase , cache_dir=_lowercase , keep_in_memory=_lowercase , streaming=_lowercase , num_proc=_lowercase , **_lowercase , ) lowerCAmelCase_ = Generator( cache_dir=_lowercase , features=_lowercase , generator=_lowercase , gen_kwargs=_lowercase , **_lowercase , ) def _lowercase ( self ) -> Tuple: '''simple docstring''' if self.streaming: lowerCAmelCase_ = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowerCAmelCase_ = None lowerCAmelCase_ = None lowerCAmelCase_ = None lowerCAmelCase_ = None self.builder.download_and_prepare( download_config=_lowercase , download_mode=_lowercase , verification_mode=_lowercase , base_path=_lowercase , num_proc=self.num_proc , ) lowerCAmelCase_ = self.builder.as_dataset( split='train' , verification_mode=_lowercase , in_memory=self.keep_in_memory ) return dataset
318
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() __lowerCAmelCase : Dict = logging.get_logger(__name__) def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Dict , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : List[str] = WavaVecaForSequenceClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : int = downstream_dict["""projector.weight"""] snake_case_ : Optional[int] = downstream_dict["""projector.bias"""] snake_case_ : List[Any] = downstream_dict["""model.post_net.linear.weight"""] snake_case_ : Union[str, Any] = downstream_dict["""model.post_net.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : Tuple , __UpperCamelCase : str ): '''simple docstring''' snake_case_ : int = WavaVecaForAudioFrameClassification.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""model.linear.weight"""] snake_case_ : int = downstream_dict["""model.linear.bias"""] return model def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[str] , __UpperCamelCase : List[Any] ): '''simple docstring''' snake_case_ : Optional[int] = WavaVecaForXVector.from_pretrained(__UpperCamelCase , config=__UpperCamelCase ) snake_case_ : Any = downstream_dict["""connector.weight"""] snake_case_ : str = downstream_dict["""connector.bias"""] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): snake_case_ : Dict = downstream_dict[ F'model.framelevel_feature_extractor.module.{i}.kernel.weight' ] snake_case_ : int = downstream_dict[F'model.framelevel_feature_extractor.module.{i}.kernel.bias'] snake_case_ : str = downstream_dict["""model.utterancelevel_feature_extractor.linear1.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear1.bias"""] snake_case_ : Any = downstream_dict["""model.utterancelevel_feature_extractor.linear2.weight"""] snake_case_ : int = downstream_dict["""model.utterancelevel_feature_extractor.linear2.bias"""] snake_case_ : List[str] = downstream_dict["""objective.W"""] return model @torch.no_grad() def __lowerCAmelCase ( __UpperCamelCase : List[Any] , __UpperCamelCase : Any , __UpperCamelCase : List[str] , __UpperCamelCase : Tuple ): '''simple docstring''' snake_case_ : Any = torch.load(__UpperCamelCase , map_location="""cpu""" ) snake_case_ : Any = checkpoint["""Downstream"""] snake_case_ : Optional[Any] = WavaVecaConfig.from_pretrained(__UpperCamelCase ) snake_case_ : Optional[int] = WavaVecaFeatureExtractor.from_pretrained( __UpperCamelCase , return_attention_mask=__UpperCamelCase , do_normalize=__UpperCamelCase ) snake_case_ : Optional[Any] = hf_config.architectures[0] if arch.endswith("""ForSequenceClassification""" ): snake_case_ : Tuple = convert_classification(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForAudioFrameClassification""" ): snake_case_ : Union[str, Any] = convert_diarization(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) elif arch.endswith("""ForXVector""" ): snake_case_ : List[str] = convert_xvector(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: raise NotImplementedError(F'S3PRL weights conversion is not supported for {arch}' ) if hf_config.use_weighted_layer_sum: snake_case_ : List[Any] = checkpoint["""Featurizer"""]["""weights"""] hf_feature_extractor.save_pretrained(__UpperCamelCase ) hf_model.save_pretrained(__UpperCamelCase ) if __name__ == "__main__": __lowerCAmelCase : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--base_model_name''', default=None, type=str, help='''Name of the huggingface pretrained base model.''' ) parser.add_argument('''--config_path''', default=None, type=str, help='''Path to the huggingface classifier config.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to the s3prl checkpoint.''') parser.add_argument('''--model_dump_path''', default=None, type=str, help='''Path to the final converted model.''') __lowerCAmelCase : Dict = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
58
0
"""simple docstring""" _a = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on _a = {value: key for key, value in MORSE_CODE_DICT.items()} def lowerCamelCase__ ( __snake_case ) -> str: """simple docstring""" return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowerCamelCase__ ( __snake_case ) -> Tuple: """simple docstring""" return "".join(REVERSE_DICT[char] for char in message.split() ) def lowerCamelCase__ ( ) -> int: """simple docstring""" _UpperCamelCase = """Morse code here!""" print(__UpperCamelCase ) _UpperCamelCase = encrypt(__UpperCamelCase ) print(__UpperCamelCase ) _UpperCamelCase = decrypt(__UpperCamelCase ) print(__UpperCamelCase ) if __name__ == "__main__": main()
19
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __lowerCAmelCase : Tuple = logging.get_logger(__name__) __lowerCAmelCase : int = {'''vocab_file''': '''vocab.txt'''} __lowerCAmelCase : Union[str, Any] = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } __lowerCAmelCase : Optional[Any] = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } __lowerCAmelCase : Any = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = PRETRAINED_INIT_CONFIGURATION _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase = ConvBertTokenizer def __init__( self , _lowercase=None , _lowercase=None , _lowercase=True , _lowercase="[UNK]" , _lowercase="[SEP]" , _lowercase="[PAD]" , _lowercase="[CLS]" , _lowercase="[MASK]" , _lowercase=True , _lowercase=None , **_lowercase , ) -> Union[str, Any]: '''simple docstring''' super().__init__( _lowercase , tokenizer_file=_lowercase , do_lower_case=_lowercase , unk_token=_lowercase , sep_token=_lowercase , pad_token=_lowercase , cls_token=_lowercase , mask_token=_lowercase , tokenize_chinese_chars=_lowercase , strip_accents=_lowercase , **_lowercase , ) snake_case_ : str = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _lowercase ) != do_lower_case or normalizer_state.get("""strip_accents""" , _lowercase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _lowercase ) != tokenize_chinese_chars ): snake_case_ : Optional[int] = getattr(_lowercase , normalizer_state.pop("""type""" ) ) snake_case_ : Dict = do_lower_case snake_case_ : str = strip_accents snake_case_ : Optional[Any] = tokenize_chinese_chars snake_case_ : int = normalizer_class(**_lowercase ) snake_case_ : Optional[int] = do_lower_case def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> int: '''simple docstring''' snake_case_ : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> List[int]: '''simple docstring''' snake_case_ : int = [self.sep_token_id] snake_case_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase__ ( self , _lowercase , _lowercase = None ) -> Tuple[str]: '''simple docstring''' snake_case_ : Optional[int] = self._tokenizer.model.save(_lowercase , name=_lowercase ) return tuple(_lowercase )
58
0
'''simple docstring''' def a__ ( a__ ): """simple docstring""" assert isinstance(__UpperCamelCase , __UpperCamelCase ), F'The input value of [n={number}] is not an integer' if number == 1: return 2 elif number < 1: __SCREAMING_SNAKE_CASE = F'The input value of [n={number}] has to be > 0' raise ValueError(__UpperCamelCase ) else: __SCREAMING_SNAKE_CASE = sylvester(number - 1 ) __SCREAMING_SNAKE_CASE = num - 1 __SCREAMING_SNAKE_CASE = num return lower * upper + 1 if __name__ == "__main__": print(f"""The 8th number in Sylvester\'s sequence: {sylvester(8)}""")
627
"""simple docstring""" import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" @register_to_config def __init__( self , _lowercase = 1_2_8 , _lowercase = 2_5_6 , _lowercase = 2000.0 , _lowercase = 7_6_8 , _lowercase = 1_2 , _lowercase = 1_2 , _lowercase = 6_4 , _lowercase = 2_0_4_8 , _lowercase = 0.1 , ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Optional[Any] = nn.Sequential( nn.Linear(_lowercase , d_model * 4 , bias=_lowercase ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=_lowercase ) , nn.SiLU() , ) snake_case_ : Any = nn.Embedding(_lowercase , _lowercase ) snake_case_ : Union[str, Any] = False snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Union[str, Any] = nn.Dropout(p=_lowercase ) snake_case_ : Tuple = nn.ModuleList() for lyr_num in range(_lowercase ): # FiLM conditional T5 decoder snake_case_ : Union[str, Any] = DecoderLayer(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) self.decoders.append(_lowercase ) snake_case_ : List[Any] = TaLayerNorm(_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(p=_lowercase ) snake_case_ : List[Any] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[int] = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def UpperCAmelCase__ ( self , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ , snake_case_ , snake_case_ : str = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. snake_case_ : Optional[int] = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) snake_case_ : int = self.conditioning_emb(_lowercase ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) snake_case_ : Tuple = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. snake_case_ : Dict = torch.broadcast_to( torch.arange(_lowercase , device=decoder_input_tokens.device ) , (batch, seq_length) , ) snake_case_ : Tuple = self.position_encoding(_lowercase ) snake_case_ : Optional[Any] = self.continuous_inputs_projection(_lowercase ) inputs += position_encodings snake_case_ : List[Any] = self.dropout(_lowercase ) # decoder: No padding present. snake_case_ : Tuple = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. snake_case_ : int = [(x, self.encoder_decoder_mask(_lowercase , _lowercase )) for x, y in encodings_and_masks] # cross attend style: concat encodings snake_case_ : Optional[Any] = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) snake_case_ : str = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: snake_case_ : int = lyr( _lowercase , conditioning_emb=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , )[0] snake_case_ : int = self.decoder_norm(_lowercase ) snake_case_ : Union[str, Any] = self.post_dropout(_lowercase ) snake_case_ : int = self.spec_out(_lowercase ) return spec_out class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase=1E-6 ) -> Union[str, Any]: '''simple docstring''' super().__init__() snake_case_ : Any = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=_lowercase , d_kv=_lowercase , num_heads=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase , layer_norm_epsilon=_lowercase ) ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , _lowercase=None , ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = self.layer[0]( _lowercase , conditioning_emb=_lowercase , attention_mask=_lowercase , ) if encoder_hidden_states is not None: snake_case_ : Tuple = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) snake_case_ : str = self.layer[1]( _lowercase , key_value_states=_lowercase , attention_mask=_lowercase , ) # Apply Film Conditional Feed Forward layer snake_case_ : Any = self.layer[-1](_lowercase , _lowercase ) return (hidden_states,) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> str: '''simple docstring''' super().__init__() snake_case_ : Any = TaLayerNorm(_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Union[str, Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : List[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[Any]: '''simple docstring''' snake_case_ : Dict = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : str = self.FiLMLayer(_lowercase , _lowercase ) # Self-attention block snake_case_ : List[Any] = self.attention(_lowercase ) snake_case_ : List[str] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' super().__init__() snake_case_ : List[Any] = Attention(query_dim=_lowercase , heads=_lowercase , dim_head=_lowercase , out_bias=_lowercase , scale_qk=_lowercase ) snake_case_ : Union[str, Any] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Optional[Any] = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None , _lowercase=None , ) -> Optional[int]: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) snake_case_ : Optional[Any] = self.attention( _lowercase , encoder_hidden_states=_lowercase , attention_mask=attention_mask.squeeze(1 ) , ) snake_case_ : Any = hidden_states + self.dropout(_lowercase ) return layer_output class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase , _lowercase ) -> Dict: '''simple docstring''' super().__init__() snake_case_ : Tuple = TaDenseGatedActDense(d_model=_lowercase , d_ff=_lowercase , dropout_rate=_lowercase ) snake_case_ : List[Any] = TaFiLMLayer(in_features=d_model * 4 , out_features=_lowercase ) snake_case_ : Optional[int] = TaLayerNorm(_lowercase , eps=_lowercase ) snake_case_ : Tuple = nn.Dropout(_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase=None ) -> str: '''simple docstring''' snake_case_ : List[Any] = self.layer_norm(_lowercase ) if conditioning_emb is not None: snake_case_ : Optional[int] = self.film(_lowercase , _lowercase ) snake_case_ : int = self.DenseReluDense(_lowercase ) snake_case_ : Optional[Any] = hidden_states + self.dropout(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase , _lowercase ) -> Optional[int]: '''simple docstring''' super().__init__() snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Optional[int] = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : Any = nn.Linear(_lowercase , _lowercase , bias=_lowercase ) snake_case_ : int = nn.Dropout(_lowercase ) snake_case_ : Optional[int] = NewGELUActivation() def UpperCAmelCase__ ( self , _lowercase ) -> int: '''simple docstring''' snake_case_ : str = self.act(self.wi_a(_lowercase ) ) snake_case_ : Dict = self.wi_a(_lowercase ) snake_case_ : Any = hidden_gelu * hidden_linear snake_case_ : List[Any] = self.dropout(_lowercase ) snake_case_ : Tuple = self.wo(_lowercase ) return hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1E-6 ) -> str: '''simple docstring''' super().__init__() snake_case_ : Union[str, Any] = nn.Parameter(torch.ones(_lowercase ) ) snake_case_ : int = eps def UpperCAmelCase__ ( self , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : Tuple = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=_lowercase ) snake_case_ : Any = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: snake_case_ : str = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class _lowerCAmelCase ( nn.Module ): """simple docstring""" def UpperCAmelCase__ ( self , _lowercase ) -> torch.Tensor: '''simple docstring''' return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.04_4715 * torch.pow(_lowercase , 3.0 )) )) class _lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self , _lowercase , _lowercase ) -> Any: '''simple docstring''' super().__init__() snake_case_ : List[Any] = nn.Linear(_lowercase , out_features * 2 , bias=_lowercase ) def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' snake_case_ : List[Any] = self.scale_bias(_lowercase ) snake_case_ , snake_case_ : Any = torch.chunk(_lowercase , 2 , -1 ) snake_case_ : Optional[Any] = x * (1 + scale) + shift return x
58
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ = logging.get_logger(__name__) snake_case_ = { '''BridgeTower/bridgetower-base''': '''https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json''', '''BridgeTower/bridgetower-base-itm-mlm''': ( '''https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): A_ : Any = 'bridgetower_vision_model' def __init__(self : List[str] , a__ : Optional[int]=768 , a__ : str=12 , a__ : Dict=3 , a__ : List[str]=16 , a__ : str=288 , a__ : List[Any]=1 , a__ : Any=1E-05 , a__ : Tuple=False , a__ : str=True , a__ : Tuple=False , **a__ : Optional[int] , ): """simple docstring""" super().__init__(**_lowercase ) __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_channels __snake_case = patch_size __snake_case = image_size __snake_case = initializer_factor __snake_case = layer_norm_eps __snake_case = stop_gradient __snake_case = share_layernorm __snake_case = remove_last_layer @classmethod def a (cls : Dict , a__ : Union[str, Any] , **a__ : List[Any] ): """simple docstring""" __snake_case = cls.get_config_dict(_lowercase , **_lowercase ) if config_dict.get('''model_type''' ) == "bridgetower": __snake_case = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(_lowercase , **_lowercase ) class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): A_ : Optional[Any] = 'bridgetower_text_model' def __init__(self : List[Any] , a__ : Tuple=5_0265 , a__ : str=768 , a__ : Any=12 , a__ : Any=12 , a__ : Optional[int]=1 , a__ : str=3072 , a__ : Any="gelu" , a__ : str=0.1 , a__ : Union[str, Any]=0.1 , a__ : str=514 , a__ : Union[str, Any]=1 , a__ : Optional[Any]=1E-05 , a__ : str=1 , a__ : Dict=0 , a__ : Tuple=2 , a__ : str="absolute" , a__ : Tuple=True , **a__ : List[Any] , ): """simple docstring""" super().__init__(**_lowercase ) __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = hidden_act __snake_case = initializer_factor __snake_case = intermediate_size __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = layer_norm_eps __snake_case = position_embedding_type __snake_case = use_cache __snake_case = pad_token_id __snake_case = bos_token_id __snake_case = eos_token_id @classmethod def a (cls : List[str] , a__ : str , **a__ : str ): """simple docstring""" __snake_case = cls.get_config_dict(_lowercase , **_lowercase ) if config_dict.get('''model_type''' ) == "bridgetower": __snake_case = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(_lowercase , **_lowercase ) class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): A_ : List[Any] = 'bridgetower' def __init__(self : Optional[int] , a__ : List[str]=True , a__ : int="gelu" , a__ : Optional[Any]=768 , a__ : Any=1 , a__ : Any=1E-05 , a__ : Optional[int]=False , a__ : Tuple="add" , a__ : List[Any]=12 , a__ : Any=6 , a__ : Dict=False , a__ : str=False , a__ : Tuple=None , a__ : int=None , **a__ : List[str] , ): """simple docstring""" __snake_case = kwargs.pop('''text_config_dict''' , _lowercase ) __snake_case = kwargs.pop('''vision_config_dict''' , _lowercase ) super().__init__(**_lowercase ) __snake_case = share_cross_modal_transformer_layers __snake_case = hidden_act __snake_case = hidden_size __snake_case = initializer_factor __snake_case = layer_norm_eps __snake_case = share_link_tower_layers __snake_case = link_tower_type __snake_case = num_attention_heads __snake_case = num_hidden_layers __snake_case = tie_word_embeddings __snake_case = init_layernorm_from_vision_encoder if text_config is None: __snake_case = {} logger.info('''`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.''' ) if vision_config is None: __snake_case = {} logger.info('''`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.''' ) __snake_case = BridgeTowerTextConfig(**_lowercase ) __snake_case = BridgeTowerVisionConfig(**_lowercase ) @classmethod def a (cls : Tuple , a__ : Union[str, Any] , a__ : Tuple , **a__ : Optional[Any] ): """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_lowercase ) def a (self : Tuple ): """simple docstring""" __snake_case = copy.deepcopy(self.__dict__ ) __snake_case = self.text_config.to_dict() __snake_case = self.vision_config.to_dict() __snake_case = self.__class__.model_type return output
592
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = '''roformer''' def __init__( self , _lowercase=5_0_0_0_0 , _lowercase=None , _lowercase=7_6_8 , _lowercase=1_2 , _lowercase=1_2 , _lowercase=3_0_7_2 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=1_5_3_6 , _lowercase=2 , _lowercase=0.02 , _lowercase=1E-12 , _lowercase=0 , _lowercase=False , _lowercase=True , **_lowercase , ) -> int: '''simple docstring''' super().__init__(pad_token_id=_lowercase , **_lowercase ) snake_case_ : str = vocab_size snake_case_ : Any = hidden_size if embedding_size is None else embedding_size snake_case_ : List[str] = hidden_size snake_case_ : str = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : Union[str, Any] = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : Tuple = initializer_range snake_case_ : str = layer_norm_eps snake_case_ : List[str] = rotary_value snake_case_ : str = use_cache class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": snake_case_ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case_ : Any = {0: """batch""", 1: """sequence"""} snake_case_ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
58
0
"""simple docstring""" from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min_version, require_tpu, require_xpu, skip, slow, ) from .training import RegressionDataset, RegressionModel, RegressionModelaXPU from .scripts import test_script, test_sync, test_ops # isort: skip
581
"""simple docstring""" import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __lowerCAmelCase ( __UpperCamelCase : Dict ): '''simple docstring''' snake_case_ : Dict = checkpoints.load_tax_checkpoint(__UpperCamelCase ) snake_case_ : Tuple = flatten_dict(__UpperCamelCase ) return flax_params def __lowerCAmelCase ( __UpperCamelCase : Optional[Any] ): '''simple docstring''' snake_case_ : Optional[Any] = {} snake_case_ : List[Any] = { """token_embedder""": """embeddings""", """encoder_norm""": """layernorm""", """kernel""": """weight""", """.out""": """.output""", """scale""": """weight""", """embedders_0.pos_embedding""": """row_embedder.weight""", """embedders_1.pos_embedding""": """column_embedder.weight""", } snake_case_ : Optional[Any] = { """query""": """attention.query""", """key""": """attention.key""", """value""": """attention.value""", """output.dense""": """output""", """encoder_decoder_attention.o""": """encoder_decoder_attention.attention.o""", """pre_self_attention_layer_norm""": """self_attention.layer_norm""", """pre_cross_attention_layer_norm""": """encoder_decoder_attention.layer_norm""", """mlp.""": """mlp.DenseReluDense.""", """pre_mlp_layer_norm""": """mlp.layer_norm""", """self_attention.o""": """self_attention.attention.o""", """decoder.embeddings.embedding""": """decoder.embed_tokens.weight""", """decoder.relpos_bias.rel_embedding""": """decoder.layer.0.self_attention.attention.relative_attention_bias.weight""", """decoder.decoder_norm.weight""": """decoder.final_layer_norm.weight""", """decoder.logits_dense.weight""": """decoder.lm_head.weight""", } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key snake_case_ : List[Any] = """.""".join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): snake_case_ : List[str] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): snake_case_ : Optional[int] = new_key.replace(__UpperCamelCase , __UpperCamelCase ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number snake_case_ : Optional[Any] = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Union[str, Any] = new_key.replace("""encoder""" , """encoder.encoder""" ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number snake_case_ : int = re.sub(r"""layers_(\d+)""" , r"""layer.\1""" , __UpperCamelCase ) snake_case_ : Dict = flax_dict[key] snake_case_ : Tuple = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): snake_case_ : Optional[int] = torch.from_numpy(converted_dict[key].T ) else: snake_case_ : List[Any] = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __lowerCAmelCase ( __UpperCamelCase : Tuple , __UpperCamelCase : List[Any] , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : List[str]=False ): '''simple docstring''' snake_case_ : Optional[int] = get_flax_param(__UpperCamelCase ) if not use_large: snake_case_ : Optional[int] = PixaStructVisionConfig() snake_case_ : Optional[Any] = PixaStructTextConfig() else: snake_case_ : Tuple = PixaStructVisionConfig( hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_attention_heads=2_4 , num_hidden_layers=1_8 ) snake_case_ : List[str] = PixaStructTextConfig(hidden_size=1_5_3_6 , d_ff=3_9_6_8 , num_heads=2_4 , num_layers=1_8 ) snake_case_ : str = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=__UpperCamelCase ) snake_case_ : Optional[int] = PixaStructForConditionalGeneration(__UpperCamelCase ) snake_case_ : str = rename_and_convert_flax_params(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) snake_case_ : Optional[int] = AutoTokenizer.from_pretrained("""ybelkada/test-pix2struct-tokenizer""" ) snake_case_ : int = PixaStructImageProcessor() snake_case_ : str = PixaStructProcessor(image_processor=__UpperCamelCase , tokenizer=__UpperCamelCase ) if use_large: snake_case_ : Optional[Any] = 4_0_9_6 snake_case_ : int = True # mkdir if needed os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) processor.save_pretrained(__UpperCamelCase ) print("""Model saved in {}""".format(__UpperCamelCase ) ) if __name__ == "__main__": __lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') __lowerCAmelCase : List[Any] = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
58
0
def lowerCAmelCase__( lowercase : float , lowercase : float ) -> Dict: if mass < 0: raise ValueError("The mass of a body cannot be negative" ) return 0.5 * mass * abs(__UpperCamelCase ) * abs(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
243
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if mass < 0: raise ValueError("""The mass of a body cannot be negative""" ) return 0.5 * mass * abs(__UpperCamelCase ) * abs(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
58
0
"""simple docstring""" import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class __lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case = "microsoft/speecht5_tts" snake_case = ( "This is a tool that reads an English text out loud. It takes an input named `text` which should contain the " "text to read (in English) and returns a waveform object containing the sound." ) snake_case = "text_reader" snake_case = SpeechTaProcessor snake_case = SpeechTaForTextToSpeech snake_case = SpeechTaHifiGan snake_case = ["text"] snake_case = ["audio"] def lowerCamelCase__ ( self : Dict ) -> Optional[Any]: """simple docstring""" if self.post_processor is None: A_ = """microsoft/speecht5_hifigan""" super().setup() def lowerCamelCase__ ( self : Any , _snake_case : List[str] , _snake_case : int=None ) -> Optional[int]: """simple docstring""" A_ = self.pre_processor(text=_lowercase , return_tensors="pt" , truncation=_lowercase ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError("Datasets needs to be installed if not passing speaker embeddings." ) A_ = load_dataset("Matthijs/cmu-arctic-xvectors" , split="validation" ) A_ = torch.tensor(embeddings_dataset[7_305]["xvector"] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def lowerCamelCase__ ( self : str , _snake_case : int ) -> Any: """simple docstring""" with torch.no_grad(): return self.model.generate_speech(**_lowercase ) def lowerCamelCase__ ( self : Tuple , _snake_case : Optional[Any] ) -> str: """simple docstring""" with torch.no_grad(): return self.post_processor(_lowercase ).cpu().detach()
115
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = StableDiffusionInpaintPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS _lowerCamelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowerCamelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase = frozenset([] ) def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) snake_case_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=_lowercase , ) snake_case_ : Dict = PNDMScheduler(skip_prk_steps=_lowercase ) torch.manual_seed(0 ) snake_case_ : str = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) snake_case_ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) snake_case_ : Dict = CLIPTextModel(_lowercase ) snake_case_ : Optional[Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case_ : int = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self , _lowercase , _lowercase=0 ) -> Optional[Any]: '''simple docstring''' snake_case_ : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowercase ) ).to(_lowercase ) snake_case_ : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] snake_case_ : Tuple = Image.fromarray(np.uinta(_lowercase ) ).convert("""RGB""" ).resize((6_4, 6_4) ) snake_case_ : Any = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((6_4, 6_4) ) if str(_lowercase ).startswith("""mps""" ): snake_case_ : str = torch.manual_seed(_lowercase ) else: snake_case_ : List[str] = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) snake_case_ : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case_ : List[str] = self.get_dummy_components() snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline(**_lowercase ) snake_case_ : Dict = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) snake_case_ : Optional[int] = self.get_dummy_inputs(_lowercase ) snake_case_ : List[str] = sd_pipe(**_lowercase ).images snake_case_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) snake_case_ : Optional[int] = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self ) -> str: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' snake_case_ : Union[str, Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Dict = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained(_lowercase , safety_checker=_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : List[Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[str] = torch.manual_seed(0 ) snake_case_ : Dict = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : Tuple = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ : str = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : Union[str, Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) snake_case_ : str = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , torch_dtype=torch.floataa , safety_checker=_lowercase , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() snake_case_ : Union[str, Any] = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : Optional[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="""np""" , ) snake_case_ : str = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case_ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) snake_case_ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) snake_case_ : int = """stabilityai/stable-diffusion-2-inpainting""" snake_case_ : List[str] = PNDMScheduler.from_pretrained(_lowercase , subfolder="""scheduler""" ) snake_case_ : Optional[Any] = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , safety_checker=_lowercase , scheduler=_lowercase , torch_dtype=torch.floataa , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case_ : Any = """Face of a yellow cat, high resolution, sitting on a park bench""" snake_case_ : List[Any] = torch.manual_seed(0 ) snake_case_ : Any = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , num_inference_steps=2 , output_type="""np""" , ) snake_case_ : Dict = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
58
0
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class __UpperCamelCase : """simple docstring""" def __init__( self : List[Any] , _A : Tuple , _A : List[Any]=13 , _A : Union[str, Any]=7 , _A : int=True , _A : str=True , _A : Optional[int]=True , _A : int=True , _A : Any=99 , _A : Any=[1, 1, 2] , _A : Dict=1 , _A : Tuple=32 , _A : Optional[Any]=4 , _A : int=8 , _A : Optional[Any]=37 , _A : int="gelu_new" , _A : str=0.1 , _A : Dict=0.1 , _A : Any=0.0 , _A : Union[str, Any]=512 , _A : Tuple=3 , _A : Dict=0.02 , _A : List[str]=3 , _A : Union[str, Any]=4 , _A : List[Any]=None , _A : str=False , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = parent __SCREAMING_SNAKE_CASE : Optional[int] = batch_size __SCREAMING_SNAKE_CASE : Tuple = seq_length __SCREAMING_SNAKE_CASE : List[Any] = is_training __SCREAMING_SNAKE_CASE : Any = use_input_mask __SCREAMING_SNAKE_CASE : List[Any] = use_token_type_ids __SCREAMING_SNAKE_CASE : Any = use_labels __SCREAMING_SNAKE_CASE : Tuple = vocab_size __SCREAMING_SNAKE_CASE : Tuple = block_sizes __SCREAMING_SNAKE_CASE : Optional[int] = num_decoder_layers __SCREAMING_SNAKE_CASE : Union[str, Any] = d_model __SCREAMING_SNAKE_CASE : Any = n_head __SCREAMING_SNAKE_CASE : Optional[int] = d_head __SCREAMING_SNAKE_CASE : Dict = d_inner __SCREAMING_SNAKE_CASE : List[Any] = hidden_act __SCREAMING_SNAKE_CASE : str = hidden_dropout __SCREAMING_SNAKE_CASE : int = attention_dropout __SCREAMING_SNAKE_CASE : List[Any] = activation_dropout __SCREAMING_SNAKE_CASE : List[str] = max_position_embeddings __SCREAMING_SNAKE_CASE : List[str] = type_vocab_size __SCREAMING_SNAKE_CASE : Tuple = 2 __SCREAMING_SNAKE_CASE : Any = num_labels __SCREAMING_SNAKE_CASE : int = num_choices __SCREAMING_SNAKE_CASE : Union[str, Any] = scope __SCREAMING_SNAKE_CASE : int = initializer_std # Used in the tests to check the size of the first attention layer __SCREAMING_SNAKE_CASE : Union[str, Any] = n_head # Used in the tests to check the size of the first hidden state __SCREAMING_SNAKE_CASE : List[str] = self.d_model # Used in the tests to check the number of output hidden states/attentions __SCREAMING_SNAKE_CASE : Tuple = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __SCREAMING_SNAKE_CASE : int = self.num_hidden_layers + 2 def UpperCAmelCase__ ( self : Optional[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE : int = None if self.use_input_mask: __SCREAMING_SNAKE_CASE : int = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE : int = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE : Dict = None __SCREAMING_SNAKE_CASE : Dict = None __SCREAMING_SNAKE_CASE : Optional[Any] = None if self.use_labels: __SCREAMING_SNAKE_CASE : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __SCREAMING_SNAKE_CASE : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __SCREAMING_SNAKE_CASE : int = ids_tensor([self.batch_size] , self.num_choices ) __SCREAMING_SNAKE_CASE : Union[str, Any] = FunnelConfig( vocab_size=self.vocab_size , block_sizes=self.block_sizes , num_decoder_layers=self.num_decoder_layers , d_model=self.d_model , n_head=self.n_head , d_head=self.d_head , d_inner=self.d_inner , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , activation_dropout=self.activation_dropout , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_std=self.initializer_std , ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def UpperCAmelCase__ ( self : List[Any] , _A : Union[str, Any] , _A : Optional[Any] , _A : Dict , _A : Dict , _A : int , _A : Optional[int] , _A : Dict , ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = TFFunnelModel(config=_lowercase ) __SCREAMING_SNAKE_CASE : str = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : Dict = model(_lowercase ) __SCREAMING_SNAKE_CASE : List[Any] = [input_ids, input_mask] __SCREAMING_SNAKE_CASE : List[str] = model(_lowercase ) __SCREAMING_SNAKE_CASE : List[Any] = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __SCREAMING_SNAKE_CASE : Optional[int] = False __SCREAMING_SNAKE_CASE : List[str] = TFFunnelModel(config=_lowercase ) __SCREAMING_SNAKE_CASE : int = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) __SCREAMING_SNAKE_CASE : List[Any] = False __SCREAMING_SNAKE_CASE : Optional[int] = TFFunnelModel(config=_lowercase ) __SCREAMING_SNAKE_CASE : Optional[Any] = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.d_model) ) def UpperCAmelCase__ ( self : Union[str, Any] , _A : List[Any] , _A : List[str] , _A : str , _A : Any , _A : List[str] , _A : int , _A : List[str] , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Tuple = TFFunnelBaseModel(config=_lowercase ) __SCREAMING_SNAKE_CASE : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : Tuple = model(_lowercase ) __SCREAMING_SNAKE_CASE : Optional[Any] = [input_ids, input_mask] __SCREAMING_SNAKE_CASE : Any = model(_lowercase ) __SCREAMING_SNAKE_CASE : Any = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) __SCREAMING_SNAKE_CASE : List[Any] = False __SCREAMING_SNAKE_CASE : List[str] = TFFunnelBaseModel(config=_lowercase ) __SCREAMING_SNAKE_CASE : str = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 3, self.d_model) ) __SCREAMING_SNAKE_CASE : List[str] = False __SCREAMING_SNAKE_CASE : Any = TFFunnelBaseModel(config=_lowercase ) __SCREAMING_SNAKE_CASE : List[str] = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, 2, self.d_model) ) def UpperCAmelCase__ ( self : Optional[int] , _A : str , _A : List[str] , _A : List[str] , _A : Any , _A : str , _A : Any , _A : Optional[int] , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = TFFunnelForPreTraining(config=_lowercase ) __SCREAMING_SNAKE_CASE : Optional[int] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : List[Any] = model(_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase__ ( self : List[str] , _A : int , _A : Dict , _A : Dict , _A : List[str] , _A : Any , _A : int , _A : Union[str, Any] , ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = TFFunnelForMaskedLM(config=_lowercase ) __SCREAMING_SNAKE_CASE : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : Optional[int] = model(_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase__ ( self : List[Any] , _A : Tuple , _A : int , _A : Dict , _A : Dict , _A : List[Any] , _A : Optional[int] , _A : Optional[int] , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = self.num_labels __SCREAMING_SNAKE_CASE : List[str] = TFFunnelForSequenceClassification(config=_lowercase ) __SCREAMING_SNAKE_CASE : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : Union[str, Any] = model(_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase__ ( self : Any , _A : Any , _A : List[Any] , _A : List[Any] , _A : Optional[int] , _A : int , _A : Tuple , _A : int , ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = self.num_choices __SCREAMING_SNAKE_CASE : List[str] = TFFunnelForMultipleChoice(config=_lowercase ) __SCREAMING_SNAKE_CASE : Union[str, Any] = tf.tile(tf.expand_dims(_lowercase , 1 ) , (1, self.num_choices, 1) ) __SCREAMING_SNAKE_CASE : Optional[Any] = tf.tile(tf.expand_dims(_lowercase , 1 ) , (1, self.num_choices, 1) ) __SCREAMING_SNAKE_CASE : Union[str, Any] = tf.tile(tf.expand_dims(_lowercase , 1 ) , (1, self.num_choices, 1) ) __SCREAMING_SNAKE_CASE : int = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } __SCREAMING_SNAKE_CASE : str = model(_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase__ ( self : Tuple , _A : List[str] , _A : Dict , _A : Union[str, Any] , _A : List[Any] , _A : Union[str, Any] , _A : int , _A : List[Any] , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = self.num_labels __SCREAMING_SNAKE_CASE : int = TFFunnelForTokenClassification(config=_lowercase ) __SCREAMING_SNAKE_CASE : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : Dict = model(_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase__ ( self : str , _A : str , _A : Optional[Any] , _A : int , _A : Tuple , _A : str , _A : Tuple , _A : str , ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = TFFunnelForQuestionAnswering(config=_lowercase ) __SCREAMING_SNAKE_CASE : Optional[int] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} __SCREAMING_SNAKE_CASE : Union[str, Any] = model(_lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase__ ( self : Optional[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = self.prepare_config_and_inputs() ( __SCREAMING_SNAKE_CASE ) : Union[str, Any] = config_and_inputs __SCREAMING_SNAKE_CASE : Optional[Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase_ = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) lowerCAmelCase_ = ( { '''feature-extraction''': (TFFunnelBaseModel, TFFunnelModel), '''fill-mask''': TFFunnelForMaskedLM, '''question-answering''': TFFunnelForQuestionAnswering, '''text-classification''': TFFunnelForSequenceClassification, '''token-classification''': TFFunnelForTokenClassification, '''zero-shot''': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = TFFunnelModelTester(self ) __SCREAMING_SNAKE_CASE : Dict = ConfigTester(self , config_class=_lowercase ) def UpperCAmelCase__ ( self : int ): """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_lowercase ) def UpperCAmelCase__ ( self : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowercase ) def UpperCAmelCase__ ( self : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowercase ) def UpperCAmelCase__ ( self : Tuple ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowercase ) @require_tf class __UpperCamelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" lowerCAmelCase_ = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) lowerCAmelCase_ = False lowerCAmelCase_ = False def UpperCAmelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = TFFunnelModelTester(self , base=_lowercase ) __SCREAMING_SNAKE_CASE : List[str] = ConfigTester(self , config_class=_lowercase ) def UpperCAmelCase__ ( self : Union[str, Any] ): """simple docstring""" self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*_lowercase ) def UpperCAmelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowercase ) def UpperCAmelCase__ ( self : Optional[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_lowercase )
74
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase ( __UpperCamelCase : int , __UpperCamelCase : str , __UpperCamelCase : Any ): '''simple docstring''' snake_case_ : Optional[Any] = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] snake_case_ : Optional[int] = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } snake_case_ : Optional[Any] = F'{src_lang}-{tgt_lang}' snake_case_ : Dict = F'\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = "facebook/wmt19-{src_lang}-{tgt_lang}"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = "{texts[src_lang]}"\ninput_ids = tokenizer.encode(input, return_tensors="pt")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) snake_case_ : List[str] = os.path.join(__UpperCamelCase , """README.md""" ) print(F'Generating {path}' ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(__UpperCamelCase ) # make sure we are under the root of the project __lowerCAmelCase : str = Path(__file__).resolve().parent.parent.parent __lowerCAmelCase : Optional[int] = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[int] = model_name.split('''-''') __lowerCAmelCase : Optional[int] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
58
0
'''simple docstring''' from jiwer import compute_measures import datasets a_ : Tuple = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' a_ : Union[str, Any] = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' a_ : Optional[int] = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def lowercase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { '''predictions''': datasets.Value('''string''', id='''sequence''' ), '''references''': datasets.Value('''string''', id='''sequence''' ), } ), codebase_urls=['''https://github.com/jitsi/jiwer/'''], reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ], ) def lowercase__ ( self, lowerCAmelCase=None, lowerCAmelCase=None, lowerCAmelCase=False ): """simple docstring""" if concatenate_texts: return compute_measures(_lowercase, _lowercase )["wer"] else: lowerCamelCase_ =0 lowerCamelCase_ =0 for prediction, reference in zip(_lowercase, _lowercase ): lowerCamelCase_ =compute_measures(_lowercase, _lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
676
"""simple docstring""" __lowerCAmelCase : Tuple = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowerCAmelCase : Tuple = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __lowerCAmelCase : Any = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
58
0
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def _A ( SCREAMING_SNAKE_CASE=None ,SCREAMING_SNAKE_CASE=None ): return field(default_factory=lambda: default ,metadata=__UpperCamelCase ) @dataclass class __UpperCamelCase : '''simple docstring''' __magic_name__ = field( metadata={"help": "The csv file to plot."} ,) __magic_name__ = field( default=SCREAMING_SNAKE_CASE__ ,metadata={"help": "Whether to plot along batch size or sequence length. Defaults to sequence length."} ,) __magic_name__ = field( default=SCREAMING_SNAKE_CASE__ ,metadata={"help": "Whether the csv file has time results or memory results. Defaults to memory results."} ,) __magic_name__ = field( default=SCREAMING_SNAKE_CASE__ ,metadata={"help": "Disable logarithmic scale when plotting"} ,) __magic_name__ = field( default=SCREAMING_SNAKE_CASE__ ,metadata={ "help": "Whether the csv file has training results or inference results. Defaults to inference results." } ,) __magic_name__ = field( default=SCREAMING_SNAKE_CASE__ ,metadata={"help": "Filename under which the plot will be saved. If unused no plot is saved."} ,) __magic_name__ = list_field( default=SCREAMING_SNAKE_CASE__ ,metadata={"help": "List of model names that are used instead of the ones in the csv file."} ) def _A ( SCREAMING_SNAKE_CASE ): try: int(__UpperCamelCase ) return True except ValueError: return False def _A ( SCREAMING_SNAKE_CASE ): try: float(__UpperCamelCase ) return True except ValueError: return False class __UpperCamelCase : '''simple docstring''' def __init__( self , lowerCamelCase__ ): UpperCAmelCase__: Any = args UpperCAmelCase__: int = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline="" ) as csv_file: UpperCAmelCase__: int = csv.DictReader(_lowercase ) for row in reader: UpperCAmelCase__: Union[str, Any] = row["""model"""] self.result_dict[model_name]["bsz"].append(int(row["batch_size"] ) ) self.result_dict[model_name]["seq_len"].append(int(row["sequence_length"] ) ) if can_convert_to_int(row["result"] ): # value is not None UpperCAmelCase__: str = int(row["result"] ) elif can_convert_to_float(row["result"] ): # value is not None UpperCAmelCase__: Optional[int] = float(row["result"] ) def _UpperCAmelCase ( self ): UpperCAmelCase__: List[str] = plt.subplots() UpperCAmelCase__: Dict = """Time usage""" if self.args.is_time else """Memory usage""" UpperCAmelCase__: str = title_str + """ for training""" if self.args.is_train else title_str + """ for inference""" if not self.args.no_log_scale: # set logarithm scales ax.set_xscale("log" ) ax.set_yscale("log" ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): UpperCAmelCase__: str = sorted(set(self.result_dict[model_name]["bsz"] ) ) UpperCAmelCase__: Tuple = sorted(set(self.result_dict[model_name]["seq_len"] ) ) UpperCAmelCase__: Tuple = self.result_dict[model_name]["""result"""] (UpperCAmelCase__): int = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) UpperCAmelCase__: Optional[Any] = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: UpperCAmelCase__: Optional[Any] = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=_lowercase , ) else: UpperCAmelCase__: Dict = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) (UpperCAmelCase__): Tuple = ( ("""batch_size""", """len""") if self.args.plot_along_batch else ("""in #tokens""", """bsz""") ) UpperCAmelCase__: Dict = np.asarray(_lowercase , _lowercase )[: len(_lowercase )] plt.scatter( _lowercase , _lowercase , label=F"{label_model_name} - {inner_loop_label}: {inner_loop_value}" ) plt.plot(_lowercase , _lowercase , "--" ) title_str += F" {label_model_name} vs." UpperCAmelCase__: Tuple = title_str[:-4] UpperCAmelCase__: Union[str, Any] = """Time in s""" if self.args.is_time else """Memory in MB""" # plot plt.title(_lowercase ) plt.xlabel(_lowercase ) plt.ylabel(_lowercase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def _A ( ): UpperCAmelCase__: List[Any] = HfArgumentParser(__UpperCamelCase ) UpperCAmelCase__: List[Any] = parser.parse_args_into_dataclasses()[0] UpperCAmelCase__: Dict = Plot(args=__UpperCamelCase ) plot.plot() if __name__ == "__main__": main()
113
"""simple docstring""" from jiwer import compute_measures import datasets __lowerCAmelCase : Tuple = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' __lowerCAmelCase : Union[str, Any] = '''\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. ''' __lowerCAmelCase : Optional[int] = ''' Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> wer = datasets.load_metric("wer") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self ) -> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/jitsi/jiwer/"""] , reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", ] , ) def UpperCAmelCase__ ( self , _lowercase=None , _lowercase=None , _lowercase=False ) -> Optional[Any]: '''simple docstring''' if concatenate_texts: return compute_measures(_lowercase , _lowercase )["wer"] else: snake_case_ : List[str] = 0 snake_case_ : Optional[int] = 0 for prediction, reference in zip(_lowercase , _lowercase ): snake_case_ : Optional[Any] = compute_measures(_lowercase , _lowercase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
58
0
"""simple docstring""" from collections import defaultdict from math import ceil, sqrt def __SCREAMING_SNAKE_CASE ( A_ = 1_00_00_00 , A_ = 10 ): lowerCAmelCase__ : defaultdict = defaultdict(__UpperCamelCase ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: lowerCAmelCase__ : Optional[Any] = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: lowerCAmelCase__ : List[Any] = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(__UpperCamelCase , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 10 ) if __name__ == "__main__": print(F'''{solution() = }''')
450
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , _lowercase , _lowercase=1_3 , _lowercase=3 , _lowercase=2_2_4 , _lowercase=3_0 , _lowercase=4_0_0 , _lowercase=True , _lowercase=None , _lowercase=True , _lowercase=[0.5, 0.5, 0.5] , _lowercase=[0.5, 0.5, 0.5] , ) -> Union[str, Any]: '''simple docstring''' snake_case_ : str = size if size is not None else {"""height""": 1_8, """width""": 1_8} snake_case_ : Union[str, Any] = parent snake_case_ : Optional[Any] = batch_size snake_case_ : Dict = num_channels snake_case_ : Optional[Any] = image_size snake_case_ : Optional[Any] = min_resolution snake_case_ : List[Any] = max_resolution snake_case_ : Union[str, Any] = do_resize snake_case_ : Optional[int] = size snake_case_ : Optional[Any] = do_normalize snake_case_ : int = image_mean snake_case_ : Dict = image_std def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _lowerCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ViTImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = EfficientFormerImageProcessorTester(self ) @property def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' return self.image_proc_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , """image_mean""" ) ) self.assertTrue(hasattr(_lowercase , """image_std""" ) ) self.assertTrue(hasattr(_lowercase , """do_normalize""" ) ) self.assertTrue(hasattr(_lowercase , """do_resize""" ) ) self.assertTrue(hasattr(_lowercase , """size""" ) ) def UpperCAmelCase__ ( self ) -> Tuple: '''simple docstring''' pass def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input snake_case_ : Any = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Optional[Any] = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ : Dict = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input snake_case_ : int = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : int = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input snake_case_ : Dict = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched snake_case_ : Tuple = image_processor(_lowercase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
58
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase_ = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
318
"""simple docstring""" from __future__ import annotations from random import random from typing import Generic, TypeVar __lowerCAmelCase : int = TypeVar('''KT''') __lowerCAmelCase : Union[str, Any] = TypeVar('''VT''') class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = "root" , _lowercase = None ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Tuple = key snake_case_ : Tuple = value snake_case_ : list[Node[KT, VT]] = [] def __repr__( self ) -> str: '''simple docstring''' return f'Node({self.key}: {self.value})' @property def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' return len(self.forward ) class _lowerCAmelCase ( Generic[KT, VT] ): """simple docstring""" def __init__( self , _lowercase = 0.5 , _lowercase = 1_6 ) -> int: '''simple docstring''' snake_case_ : Node[KT, VT] = Node[KT, VT]() snake_case_ : Union[str, Any] = 0 snake_case_ : Optional[int] = p snake_case_ : Any = max_level def __str__( self ) -> str: '''simple docstring''' snake_case_ : str = list(self ) if len(_lowercase ) == 0: return f'SkipList(level={self.level})' snake_case_ : List[Any] = max((len(str(_lowercase ) ) for item in items) , default=4 ) snake_case_ : str = max(_lowercase , 4 ) + 4 snake_case_ : Union[str, Any] = self.head snake_case_ : Dict = [] snake_case_ : List[str] = node.forward.copy() lines.append(f'[{node.key}]'.ljust(_lowercase , """-""" ) + """* """ * len(_lowercase ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) while len(node.forward ) != 0: snake_case_ : Optional[Any] = node.forward[0] lines.append( f'[{node.key}]'.ljust(_lowercase , """-""" ) + """ """.join(str(n.key ) if n.key == node.key else """|""" for n in forwards ) ) lines.append(""" """ * label_size + """| """ * len(_lowercase ) ) snake_case_ : List[str] = node.forward lines.append("""None""".ljust(_lowercase ) + """* """ * len(_lowercase ) ) return f'SkipList(level={self.level})\n' + "\n".join(_lowercase ) def __iter__( self ) -> Optional[int]: '''simple docstring''' snake_case_ : Dict = self.head while len(node.forward ) != 0: yield node.forward[0].key snake_case_ : Dict = node.forward[0] def UpperCAmelCase__ ( self ) -> int: '''simple docstring''' snake_case_ : Optional[int] = 1 while random() < self.p and level < self.max_level: level += 1 return level def UpperCAmelCase__ ( self , _lowercase ) -> tuple[Node[KT, VT] | None, list[Node[KT, VT]]]: '''simple docstring''' snake_case_ : Optional[Any] = [] snake_case_ : int = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: snake_case_ : List[Any] = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(_lowercase ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def UpperCAmelCase__ ( self , _lowercase ) -> Optional[int]: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: for i, update_node in enumerate(_lowercase ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: snake_case_ : List[str] = node.forward[i] else: snake_case_ : Tuple = update_node.forward[:i] def UpperCAmelCase__ ( self , _lowercase , _lowercase ) -> str: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: snake_case_ : List[Any] = value else: snake_case_ : Optional[int] = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , _lowercase ): update_vector.append(self.head ) snake_case_ : Any = level snake_case_ : Optional[int] = Node(_lowercase , _lowercase ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(_lowercase ) else: snake_case_ : Optional[Any] = new_node def UpperCAmelCase__ ( self , _lowercase ) -> VT | None: '''simple docstring''' snake_case_ , snake_case_ : Dict = self._locate_node(_lowercase ) if node is not None: return node.value return None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : List[str] = SkipList() skip_list.insert("""Key1""" , 3 ) skip_list.insert("""Key2""" , 1_2 ) skip_list.insert("""Key3""" , 4_1 ) skip_list.insert("""Key4""" , -1_9 ) snake_case_ : Optional[int] = skip_list.head snake_case_ : List[Any] = {} while node.level != 0: snake_case_ : List[str] = node.forward[0] snake_case_ : Union[str, Any] = node.value assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 1_2 assert all_values["Key3"] == 4_1 assert all_values["Key4"] == -1_9 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_0 ) skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""Key5""" , 7 ) skip_list.insert("""Key7""" , 1_0 ) skip_list.insert("""Key10""" , 5 ) skip_list.insert("""Key7""" , 7 ) skip_list.insert("""Key5""" , 5 ) skip_list.insert("""Key10""" , 1_0 ) snake_case_ : str = skip_list.head snake_case_ : str = {} while node.level != 0: snake_case_ : Optional[Any] = node.forward[0] snake_case_ : int = node.value if len(__UpperCamelCase ) != 4: print() assert len(__UpperCamelCase ) == 4 assert all_values["Key1"] == 1_2 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 1_0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : str = SkipList() assert skip_list.find("""Some key""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[Any] = SkipList() skip_list.insert("""Key2""" , 2_0 ) assert skip_list.find("""Key2""" ) == 2_0 skip_list.insert("""Some Key""" , 1_0 ) skip_list.insert("""Key2""" , 8 ) skip_list.insert("""V""" , 1_3 ) assert skip_list.find("""Y""" ) is None assert skip_list.find("""Key2""" ) == 8 assert skip_list.find("""Some Key""" ) == 1_0 assert skip_list.find("""V""" ) == 1_3 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Any = SkipList() skip_list.delete("""Some key""" ) assert len(skip_list.head.forward ) == 0 def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Tuple = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Optional[int] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""V""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) == 1_4 assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""X""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) == 1_2 assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key1""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) == 1_5 skip_list.delete("""Key2""" ) assert skip_list.find("""V""" ) is None assert skip_list.find("""X""" ) is None assert skip_list.find("""Key1""" ) is None assert skip_list.find("""Key2""" ) is None def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Union[str, Any] = SkipList() skip_list.insert("""Key1""" , 1_2 ) skip_list.insert("""V""" , 1_3 ) skip_list.insert("""X""" , 1_4_2 ) skip_list.insert("""Key2""" , 1_5 ) skip_list.delete("""X""" ) def traverse_keys(__UpperCamelCase : str ): yield node.key for forward_node in node.forward: yield from traverse_keys(__UpperCamelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __lowerCAmelCase ( ): '''simple docstring''' def is_sorted(__UpperCamelCase : List[Any] ): return all(next_item >= item for item, next_item in zip(__UpperCamelCase , lst[1:] ) ) snake_case_ : str = SkipList() for i in range(1_0 ): skip_list.insert(__UpperCamelCase , __UpperCamelCase ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(__UpperCamelCase ) ) skip_list.insert(-1_2 , -1_2 ) skip_list.insert(7_7 , 7_7 ) assert is_sorted(list(__UpperCamelCase ) ) def __lowerCAmelCase ( ): '''simple docstring''' for _ in range(1_0_0 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = SkipList() skip_list.insert(2 , """2""" ) skip_list.insert(4 , """4""" ) skip_list.insert(6 , """4""" ) skip_list.insert(4 , """5""" ) skip_list.insert(8 , """4""" ) skip_list.insert(9 , """4""" ) skip_list.delete(4 ) print(__UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
58
0
"""simple docstring""" import warnings from functools import wraps from typing import Callable def lowerCamelCase__ ( __snake_case ) -> Dict: """simple docstring""" @wraps(__UpperCamelCase ) def _inner_fn(*__snake_case, **__snake_case ): warnings.warn( (F'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.'''), __UpperCamelCase, ) return fn(*__UpperCamelCase, **__UpperCamelCase ) return _inner_fn
19
"""simple docstring""" import argparse import os import re import packaging.version __lowerCAmelCase : Optional[Any] = '''examples/''' __lowerCAmelCase : Union[str, Any] = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __lowerCAmelCase : Union[str, Any] = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } __lowerCAmelCase : List[Any] = '''README.md''' def __lowerCAmelCase ( __UpperCamelCase : Any , __UpperCamelCase : Optional[int] , __UpperCamelCase : Dict ): '''simple docstring''' with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : Any = f.read() snake_case_ , snake_case_ : Optional[int] = REPLACE_PATTERNS[pattern] snake_case_ : Union[str, Any] = replace.replace("""VERSION""" , __UpperCamelCase ) snake_case_ : List[Any] = re_pattern.sub(__UpperCamelCase , __UpperCamelCase ) with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.write(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : List[str] ): '''simple docstring''' for folder, directories, fnames in os.walk(__UpperCamelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("""research_projects""" ) if "legacy" in directories: directories.remove("""legacy""" ) for fname in fnames: if fname.endswith(""".py""" ): update_version_in_file(os.path.join(__UpperCamelCase , __UpperCamelCase ) , __UpperCamelCase , pattern="""examples""" ) def __lowerCAmelCase ( __UpperCamelCase : List[str] , __UpperCamelCase : int=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not patch: update_version_in_examples(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = """🤗 Transformers currently provides the following architectures""" snake_case_ : Union[str, Any] = """1. Want to contribute a new model?""" with open(__UpperCamelCase , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: snake_case_ : str = f.readlines() # Find the start of the list. snake_case_ : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 snake_case_ : Optional[int] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("""1.""" ): snake_case_ : Any = lines[index].replace( """https://huggingface.co/docs/diffusers/main/model_doc""" , """https://huggingface.co/docs/diffusers/model_doc""" , ) index += 1 with open(__UpperCamelCase , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' with open(REPLACE_FILES["""init"""] , """r""" ) as f: snake_case_ : Any = f.read() snake_case_ : Tuple = REPLACE_PATTERNS["""init"""][0].search(__UpperCamelCase ).groups()[0] return packaging.version.parse(__UpperCamelCase ) def __lowerCAmelCase ( __UpperCamelCase : str=False ): '''simple docstring''' snake_case_ : Union[str, Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("""Can't create a patch version from the dev branch, checkout a released version!""" ) if default_version.is_devrelease: snake_case_ : str = default_version.base_version elif patch: snake_case_ : str = F'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: snake_case_ : str = F'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. snake_case_ : int = input(F'Which version are you releasing? [{default_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Optional[int] = default_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase , patch=__UpperCamelCase ) def __lowerCAmelCase ( ): '''simple docstring''' snake_case_ : Dict = get_version() snake_case_ : str = F'{current_version.major}.{current_version.minor + 1}.0.dev0' snake_case_ : Tuple = current_version.base_version # Check with the user we got that right. snake_case_ : Optional[int] = input(F'Which version are we developing now? [{dev_version}]' ) if len(__UpperCamelCase ) == 0: snake_case_ : Dict = dev_version print(F'Updating version to {version}.' ) global_version_update(__UpperCamelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": __lowerCAmelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __lowerCAmelCase : str = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
58
0
'''simple docstring''' def a__ ( a__ , a__ ): """simple docstring""" if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
627
"""simple docstring""" def __lowerCAmelCase ( __UpperCamelCase : float , __UpperCamelCase : float ): '''simple docstring''' if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
58
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def __init__(self : Optional[Any] , a__ : List[Any] , a__ : Any=7 , a__ : Union[str, Any]=3 , a__ : str=18 , a__ : Any=30 , a__ : Optional[int]=400 , a__ : Union[str, Any]=True , a__ : Optional[int]=None , a__ : Dict=True , a__ : Any=None , a__ : Tuple=True , a__ : Union[str, Any]=[0.5, 0.5, 0.5] , a__ : Optional[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __snake_case = size if size is not None else {"""shortest_edge""": 18} __snake_case = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} __snake_case = parent __snake_case = batch_size __snake_case = num_channels __snake_case = image_size __snake_case = min_resolution __snake_case = max_resolution __snake_case = do_resize __snake_case = size __snake_case = do_center_crop __snake_case = crop_size __snake_case = do_normalize __snake_case = image_mean __snake_case = image_std def a (self : str ): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): A_ : Any = LevitImageProcessor if is_vision_available() else None def a (self : List[str] ): """simple docstring""" __snake_case = LevitImageProcessingTester(self ) @property def a (self : str ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a (self : List[str] ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowercase , '''image_mean''' ) ) self.assertTrue(hasattr(_lowercase , '''image_std''' ) ) self.assertTrue(hasattr(_lowercase , '''do_normalize''' ) ) self.assertTrue(hasattr(_lowercase , '''do_resize''' ) ) self.assertTrue(hasattr(_lowercase , '''do_center_crop''' ) ) self.assertTrue(hasattr(_lowercase , '''size''' ) ) def a (self : Any ): """simple docstring""" __snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) __snake_case = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def a (self : Optional[int] ): """simple docstring""" pass def a (self : str ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , Image.Image ) # Test not batched input __snake_case = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __snake_case = image_processing(_lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def a (self : Tuple ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase , numpify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , np.ndarray ) # Test not batched input __snake_case = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __snake_case = image_processing(_lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def a (self : Optional[int] ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowercase , torchify=_lowercase ) for image in image_inputs: self.assertIsInstance(_lowercase , torch.Tensor ) # Test not batched input __snake_case = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __snake_case = image_processing(_lowercase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
592
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCAmelCase ( __UpperCamelCase : int ): '''simple docstring''' if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise TypeError("""Undefined for non-integers""" ) elif precision < 1: raise ValueError("""Undefined for non-natural numbers""" ) snake_case_ : str = precision snake_case_ : Any = ceil(precision / 1_4 ) snake_case_ : Dict = 4_2_6_8_8_0 * Decimal(1_0_0_0_5 ).sqrt() snake_case_ : Optional[Any] = 1 snake_case_ : List[str] = 1_3_5_9_1_4_0_9 snake_case_ : Optional[int] = Decimal(__UpperCamelCase ) for k in range(1 , __UpperCamelCase ): snake_case_ : Any = factorial(6 * k ) // (factorial(3 * k ) * factorial(__UpperCamelCase ) ** 3) linear_term += 5_4_5_1_4_0_1_3_4 exponential_term *= -2_6_2_5_3_7_4_1_2_6_4_0_7_6_8_0_0_0 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __lowerCAmelCase : int = 50 print(F'''The first {n} digits of pi is: {pi(n)}''')
58
0