code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import evaluate import numpy as np import torch from datasets import load_dataset from PIL import Image from torchvision.transforms import ( CenterCrop, Compose, Normalize, RandomHorizontalFlip, RandomResizedCrop, Resize, ToTensor, ) import transformers from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, AutoConfig, AutoImageProcessor, AutoModelForImageClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __SCREAMING_SNAKE_CASE : Tuple = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.8.0''', '''To fix: pip install -r examples/pytorch/image-classification/requirements.txt''') __SCREAMING_SNAKE_CASE : List[Any] = list(MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING.keys()) __SCREAMING_SNAKE_CASE : Tuple = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def lowerCAmelCase_( lowercase_ : str ) -> Dict: with open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = Image.open(lowercase_ ) return im.convert('''RGB''' ) @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : Optional[str] = field( default=A__, metadata={ 'help': 'Name of a dataset from the hub (could be your own, possibly private dataset hosted on the hub).' }, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) lowercase__ : Optional[str] = field(default=A__, metadata={'help': 'A folder containing the training data.'} ) lowercase__ : Optional[str] = field(default=A__, metadata={'help': 'A folder containing the validation data.'} ) lowercase__ : Optional[float] = field( default=0.15, metadata={'help': 'Percent to split off of train for validation.'} ) lowercase__ : Optional[int] = field( default=A__, metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) }, ) lowercase__ : Optional[int] = field( default=A__, metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) }, ) def snake_case__ ( self ): if self.dataset_name is None and (self.train_dir is None and self.validation_dir is None): raise ValueError( '''You must specify either a dataset name from the hub or a train and/or validation directory.''' ) @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : str = field( default='google/vit-base-patch16-224-in21k', metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'}, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'If training from scratch, pass a model type from the list: ' + ', '.join(A__ )}, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Where do you want to store the pretrained models downloaded from s3'} ) lowercase__ : str = field( default='main', metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'}, ) lowercase__ : str = field(default=A__, metadata={'help': 'Name or path of preprocessor config.'} ) lowercase__ : bool = field( default=A__, metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) }, ) lowercase__ : bool = field( default=A__, metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'}, ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Tuple: _lowerCamelCase = torch.stack([example['''pixel_values'''] for example in examples] ) _lowerCamelCase = torch.tensor([example['''labels'''] for example in examples] ) return {"pixel_values": pixel_values, "labels": labels} def lowerCAmelCase_( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowerCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_image_classification''' , lowercase_ , lowercase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowerCamelCase = training_args.get_process_log_level() logger.setLevel(lowercase_ ) transformers.utils.logging.set_verbosity(lowercase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. _lowerCamelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCamelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Initialize our dataset and prepare it for the 'image-classification' task. if data_args.dataset_name is not None: _lowerCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir , task='''image-classification''' , use_auth_token=True if model_args.use_auth_token else None , ) else: _lowerCamelCase = {} if data_args.train_dir is not None: _lowerCamelCase = os.path.join(data_args.train_dir , '''**''' ) if data_args.validation_dir is not None: _lowerCamelCase = os.path.join(data_args.validation_dir , '''**''' ) _lowerCamelCase = load_dataset( '''imagefolder''' , data_files=lowercase_ , cache_dir=model_args.cache_dir , task='''image-classification''' , ) # If we don't have a validation split, split off a percentage of train as validation. _lowerCamelCase = None if '''validation''' in dataset.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , lowercase_ ) and data_args.train_val_split > 0.0: _lowerCamelCase = dataset['''train'''].train_test_split(data_args.train_val_split ) _lowerCamelCase = split['''train'''] _lowerCamelCase = split['''test'''] # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _lowerCamelCase = dataset['''train'''].features['''labels'''].names _lowerCamelCase , _lowerCamelCase = {}, {} for i, label in enumerate(lowercase_ ): _lowerCamelCase = str(lowercase_ ) _lowerCamelCase = label # Load the accuracy metric from the datasets package _lowerCamelCase = evaluate.load('''accuracy''' ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowercase_ : Union[str, Any] ): return metric.compute(predictions=np.argmax(p.predictions , axis=1 ) , references=p.label_ids ) _lowerCamelCase = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(lowercase_ ) , labelaid=lowercase_ , idalabel=lowercase_ , finetuning_task='''image-classification''' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase = AutoModelForImageClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowercase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) _lowerCamelCase = AutoImageProcessor.from_pretrained( model_args.image_processor_name or model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Define torchvision transforms to be applied to each image. if "shortest_edge" in image_processor.size: _lowerCamelCase = image_processor.size['''shortest_edge'''] else: _lowerCamelCase = (image_processor.size['''height'''], image_processor.size['''width''']) _lowerCamelCase = Normalize(mean=image_processor.image_mean , std=image_processor.image_std ) _lowerCamelCase = Compose( [ RandomResizedCrop(lowercase_ ), RandomHorizontalFlip(), ToTensor(), normalize, ] ) _lowerCamelCase = Compose( [ Resize(lowercase_ ), CenterCrop(lowercase_ ), ToTensor(), normalize, ] ) def train_transforms(lowercase_ : Optional[Any] ): _lowerCamelCase = [ _train_transforms(pil_img.convert('''RGB''' ) ) for pil_img in example_batch['''image'''] ] return example_batch def val_transforms(lowercase_ : Tuple ): _lowerCamelCase = [_val_transforms(pil_img.convert('''RGB''' ) ) for pil_img in example_batch['''image''']] return example_batch if training_args.do_train: if "train" not in dataset: raise ValueError('''--do_train requires a train dataset''' ) if data_args.max_train_samples is not None: _lowerCamelCase = ( dataset['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms dataset["train"].set_transform(lowercase_ ) if training_args.do_eval: if "validation" not in dataset: raise ValueError('''--do_eval requires a validation dataset''' ) if data_args.max_eval_samples is not None: _lowerCamelCase = ( dataset['''validation'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms dataset["validation"].set_transform(lowercase_ ) # Initalize our trainer _lowerCamelCase = Trainer( model=lowercase_ , args=lowercase_ , train_dataset=dataset['''train'''] if training_args.do_train else None , eval_dataset=dataset['''validation'''] if training_args.do_eval else None , compute_metrics=lowercase_ , tokenizer=lowercase_ , data_collator=lowercase_ , ) # Training if training_args.do_train: _lowerCamelCase = None if training_args.resume_from_checkpoint is not None: _lowerCamelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCamelCase = last_checkpoint _lowerCamelCase = trainer.train(resume_from_checkpoint=lowercase_ ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _lowerCamelCase = trainer.evaluate() trainer.log_metrics('''eval''' , lowercase_ ) trainer.save_metrics('''eval''' , lowercase_ ) # Write model card and (optionally) push to hub _lowerCamelCase = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''image-classification''', '''dataset''': data_args.dataset_name, '''tags''': ['''image-classification''', '''vision'''], } if training_args.push_to_hub: trainer.push_to_hub(**lowercase_ ) else: trainer.create_model_card(**lowercase_ ) if __name__ == "__main__": main()
623
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[str] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/config.json''', # See all XGLM models at https://huggingface.co/models?filter=xglm } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Tuple = 'xglm' lowercase__ : Tuple = ['past_key_values'] lowercase__ : Optional[int] = { 'num_attention_heads': 'attention_heads', 'hidden_size': 'd_model', 'num_hidden_layers': 'num_layers', } def __init__( self , lowerCamelCase__=2_5_6_0_0_8 , lowerCamelCase__=2_0_4_8 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=4_0_9_6 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , **lowerCamelCase__ , ): _lowerCamelCase = vocab_size _lowerCamelCase = max_position_embeddings _lowerCamelCase = d_model _lowerCamelCase = ffn_dim _lowerCamelCase = num_layers _lowerCamelCase = attention_heads _lowerCamelCase = activation_function _lowerCamelCase = dropout _lowerCamelCase = attention_dropout _lowerCamelCase = activation_dropout _lowerCamelCase = layerdrop _lowerCamelCase = init_std _lowerCamelCase = scale_embedding # scale factor will be sqrt(d_model) if True _lowerCamelCase = use_cache super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
623
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : int ) -> str: _lowerCamelCase = [[] for _ in range(lowercase_ )] _lowerCamelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1 or len(lowercase_ ) <= key: return input_string for position, character in enumerate(lowercase_ ): _lowerCamelCase = position % (lowest * 2) # puts it in bounds _lowerCamelCase = min(lowercase_ , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(lowercase_ ) _lowerCamelCase = [''''''.join(lowercase_ ) for row in temp_grid] _lowerCamelCase = ''''''.join(lowercase_ ) return output_string def lowerCAmelCase_( lowercase_ : str , lowercase_ : int ) -> str: _lowerCamelCase = [] _lowerCamelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1: return input_string _lowerCamelCase = [[] for _ in range(lowercase_ )] # generates template for position in range(len(lowercase_ ) ): _lowerCamelCase = position % (lowest * 2) # puts it in bounds _lowerCamelCase = min(lowercase_ , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append('''*''' ) _lowerCamelCase = 0 for row in temp_grid: # fills in the characters _lowerCamelCase = input_string[counter : counter + len(lowercase_ )] grid.append(list(lowercase_ ) ) counter += len(lowercase_ ) _lowerCamelCase = '''''' # reads as zigzag for position in range(len(lowercase_ ) ): _lowerCamelCase = position % (lowest * 2) # puts it in bounds _lowerCamelCase = min(lowercase_ , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def lowerCAmelCase_( lowercase_ : str ) -> dict[int, str]: _lowerCamelCase = {} for key_guess in range(1 , len(lowercase_ ) ): # tries every key _lowerCamelCase = decrypt(lowercase_ , lowercase_ ) return results if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
1
"""simple docstring""" import math def lowerCAmelCase_( lowercase_ : int ) -> int: if not isinstance(lowercase_ , lowercase_ ): _lowerCamelCase = F"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase_ ) if number < 1: _lowerCamelCase = F"""Input value of [number={number}] must be > 0""" raise ValueError(lowercase_ ) elif number == 1: return 3 elif number == 2: return 5 else: _lowerCamelCase = int(math.log(number // 3 , 2 ) ) + 2 _lowerCamelCase = [3, 5] _lowerCamelCase = 2 _lowerCamelCase = 3 for block in range(1 , lowercase_ ): for _ in range(lowercase_ ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(1_1): __SCREAMING_SNAKE_CASE : Dict = 0 try: __SCREAMING_SNAKE_CASE : List[str] = proth(number) except ValueError: print(F"""ValueError: there is no {number}th Proth number""") continue print(F"""The {number}th Proth number: {value}""")
623
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
1
"""simple docstring""" import unittest from transformers import DonutProcessor __SCREAMING_SNAKE_CASE : Dict = '''naver-clova-ix/donut-base''' class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = DonutProcessor.from_pretrained(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = { '''name''': '''John Doe''', '''age''': '''99''', '''city''': '''Atlanta''', '''state''': '''GA''', '''zip''': '''30301''', '''phone''': '''123-4567''', '''nicknames''': [{'''nickname''': '''Johnny'''}, {'''nickname''': '''JD'''}], } _lowerCamelCase = ( '''<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>''' '''<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>''' '''<s_nicknames><s_nickname>Johnny</s_nickname>''' '''<sep/><s_nickname>JD</s_nickname></s_nicknames>''' ) _lowerCamelCase = self.processor.tokenajson(lowerCamelCase__ ) self.assertDictEqual(lowerCamelCase__ , lowerCamelCase__ )
623
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
1
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device __SCREAMING_SNAKE_CASE : Any = False class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' pass @nightly @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self ): _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger ''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = generator.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def snake_case__ ( self ): _lowerCamelCase = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger ''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=5_0 , output_type='''numpy''' ).images _lowerCamelCase = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=1_6 , lowerCamelCase__=[1, 2, 1] , lowerCamelCase__=[2, 2, 4] , lowerCamelCase__=2 , lowerCamelCase__=2.0 , lowerCamelCase__=True , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.1 , lowerCamelCase__="gelu" , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-5 , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=1_0 , lowerCamelCase__=8 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = embed_dim _lowerCamelCase = depths _lowerCamelCase = num_heads _lowerCamelCase = window_size _lowerCamelCase = mlp_ratio _lowerCamelCase = qkv_bias _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = drop_path_rate _lowerCamelCase = hidden_act _lowerCamelCase = use_absolute_embeddings _lowerCamelCase = patch_norm _lowerCamelCase = layer_norm_eps _lowerCamelCase = initializer_range _lowerCamelCase = is_training _lowerCamelCase = scope _lowerCamelCase = use_labels _lowerCamelCase = type_sequence_label_size _lowerCamelCase = encoder_stride def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = SwinvaModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _lowerCamelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = SwinvaForMaskedImageModeling(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = SwinvaForMaskedImageModeling(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.type_sequence_label_size _lowerCamelCase = SwinvaForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowercase__ : Optional[int] = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowercase__ : Any = False lowercase__ : Tuple = False lowercase__ : Tuple = False lowercase__ : Dict = False def snake_case__ ( self ): _lowerCamelCase = SwinvaModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , embed_dim=3_7 ) def snake_case__ ( self ): self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @unittest.skip(reason='''Got `CUDA error: misaligned address` with PyTorch 2.0.0.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Swinv2 does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = True for model_class in self.all_model_classes: _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = True _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.attentions _lowerCamelCase = len(self.model_tester.depths ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _lowerCamelCase = True _lowerCamelCase = config.window_size**2 _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.attentions self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) _lowerCamelCase = len(lowerCamelCase__ ) # Check attention is always last and order is fine _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) if hasattr(self.model_tester , '''num_hidden_states_types''' ): _lowerCamelCase = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states _lowerCamelCase = 2 self.assertEqual(out_len + added_hidden_states , len(lowerCamelCase__ ) ) _lowerCamelCase = outputs.attentions self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.hidden_states _lowerCamelCase = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # Swinv2 has a different seq_length _lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _lowerCamelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) _lowerCamelCase = outputs.reshaped_hidden_states self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = reshaped_hidden_states[0].shape _lowerCamelCase = ( reshaped_hidden_states[0].view(lowerCamelCase__ , lowerCamelCase__ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = 3 _lowerCamelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _lowerCamelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _lowerCamelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _lowerCamelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = SwinvaModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = _config_zero_init(lowerCamelCase__ ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( AutoImageProcessor.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = SwinvaForImageClassification.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ).to( lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
623
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Any = { '''andreasmadsen/efficient_mlm_m0.40''': ( '''https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json''' ), } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'roberta-prelayernorm' def __init__( self , lowerCamelCase__=5_0_2_6_5 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , lowerCamelCase__="absolute" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__(pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = position_embedding_type _lowerCamelCase = use_cache _lowerCamelCase = classifier_dropout class lowerCamelCase_( A__ ): '''simple docstring''' @property def snake_case__ ( self ): if self.task == "multiple-choice": _lowerCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
623
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/config.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/config.json''' # See all FNet models at https://huggingface.co/models?filter=fnet } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = 'fnet' def __init__( self , lowerCamelCase__=3_2_0_0_0 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=4 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=False , lowerCamelCase__=5_1_2 , lowerCamelCase__=3 , lowerCamelCase__=1 , lowerCamelCase__=2 , **lowerCamelCase__ , ): super().__init__(pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = max_position_embeddings _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = type_vocab_size _lowerCamelCase = layer_norm_eps _lowerCamelCase = use_tpu_fourier_optimizations _lowerCamelCase = tpu_short_seq_length
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = DanceDiffusionPipeline lowercase__ : str = UNCONDITIONAL_AUDIO_GENERATION_PARAMS lowercase__ : int = PipelineTesterMixin.required_optional_params - { 'callback', 'latents', 'callback_steps', 'output_type', 'num_images_per_prompt', } lowercase__ : List[Any] = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS lowercase__ : str = False lowercase__ : int = False def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDModel( block_out_channels=(3_2, 3_2, 6_4) , extra_in_channels=1_6 , sample_size=5_1_2 , sample_rate=1_6_0_0_0 , in_channels=2 , out_channels=2 , flip_sin_to_cos=lowerCamelCase__ , use_timestep_embedding=lowerCamelCase__ , time_embedding_type='''fourier''' , mid_block_type='''UNetMidBlock1D''' , down_block_types=('''DownBlock1DNoSkip''', '''DownBlock1D''', '''AttnDownBlock1D''') , up_block_types=('''AttnUpBlock1D''', '''UpBlock1D''', '''UpBlock1DNoSkip''') , ) _lowerCamelCase = IPNDMScheduler() _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 4, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = DanceDiffusionPipeline(**lowerCamelCase__ ) _lowerCamelCase = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ) _lowerCamelCase = output.audios _lowerCamelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) _lowerCamelCase = np.array([-0.7_2_6_5, 1.0_0_0_0, -0.8_3_8_8, 0.1_1_7_5, 0.9_4_9_8, -1.0_0_0_0] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def snake_case__ ( self ): return super().test_save_load_local() @skip_mps def snake_case__ ( self ): return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) @skip_mps def snake_case__ ( self ): return super().test_save_load_optional_components() @skip_mps def snake_case__ ( self ): return super().test_attention_slicing_forward_pass() def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self ): _lowerCamelCase = torch_device _lowerCamelCase = DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' ) _lowerCamelCase = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe(generator=lowerCamelCase__ , num_inference_steps=1_0_0 , audio_length_in_s=4.0_9_6 ) _lowerCamelCase = output.audios _lowerCamelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _lowerCamelCase = np.array([-0.0_1_9_2, -0.0_2_3_1, -0.0_3_1_8, -0.0_0_5_9, 0.0_0_0_2, -0.0_0_2_0] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): _lowerCamelCase = torch_device _lowerCamelCase = DanceDiffusionPipeline.from_pretrained('''harmonai/maestro-150k''' , torch_dtype=torch.floataa ) _lowerCamelCase = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe(generator=lowerCamelCase__ , num_inference_steps=1_0_0 , audio_length_in_s=4.0_9_6 ) _lowerCamelCase = output.audios _lowerCamelCase = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) _lowerCamelCase = np.array([-0.0_3_6_7, -0.0_4_8_8, -0.0_7_7_1, -0.0_5_2_5, -0.0_4_4_4, -0.0_3_4_1] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('''0.12.2'''): raise Exception('''requires fairseq >= 0.12.2''') if version.parse(fairseq.__version__) > version.parse('''2'''): raise Exception('''requires fairseq < v2''') logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[Any] = '''Hello, World!''' __SCREAMING_SNAKE_CASE : Any = '''en_XX''' def lowerCAmelCase_( lowercase_ : str , lowercase_ : str , lowercase_ : bool ) -> List[str]: _lowerCamelCase = Path('''data_bin''' ) _lowerCamelCase = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(lowercase_ ).parent ) , checkpoint_file=Path(lowercase_ ).name , _name='''xmod_base''' , arch='''xmod_base''' , task='''multilingual_masked_lm''' , data_name_or_path=str(lowercase_ ) , bpe='''sentencepiece''' , sentencepiece_model=str(Path(lowercase_ ).parent / '''sentencepiece.bpe.model''' ) , src_dict=str(data_dir / '''dict.txt''' ) , ) xmod.eval() # disable dropout print(lowercase_ ) _lowerCamelCase = xmod.model.encoder.sentence_encoder _lowerCamelCase = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings , hidden_size=xmod.cfg.model.encoder_embed_dim , num_hidden_layers=xmod.cfg.model.encoder_layers , num_attention_heads=xmod.cfg.model.encoder_attention_heads , intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=5_14 , type_vocab_size=1 , layer_norm_eps=1e-5 , pre_norm=xmod.cfg.model.encoder_normalize_before , adapter_reduction_factor=getattr(xmod.cfg.model , '''bottleneck''' , 2 ) , adapter_layer_norm=xmod.cfg.model.adapter_layer_norm , adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm , ln_before_adapter=xmod.cfg.model.ln_before_adapter , languages=xmod.cfg.model.languages , ) if classification_head: _lowerCamelCase = xmod.model.classification_heads['''mnli'''].out_proj.weight.shape[0] print('''Our X-MOD config:''' , lowercase_ ) _lowerCamelCase = XmodForSequenceClassification(lowercase_ ) if classification_head else XmodForMaskedLM(lowercase_ ) model.eval() # Now let's copy all the weights. # Embeddings _lowerCamelCase = xmod_sent_encoder.embed_tokens.weight _lowerCamelCase = xmod_sent_encoder.embed_positions.weight _lowerCamelCase = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. _lowerCamelCase = xmod_sent_encoder.layernorm_embedding.weight _lowerCamelCase = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer _lowerCamelCase = model.roberta.encoder.layer[i] _lowerCamelCase = xmod_sent_encoder.layers[i] # self attention _lowerCamelCase = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError('''Dimensions of self-attention weights do not match.''' ) _lowerCamelCase = xmod_layer.self_attn.q_proj.weight _lowerCamelCase = xmod_layer.self_attn.q_proj.bias _lowerCamelCase = xmod_layer.self_attn.k_proj.weight _lowerCamelCase = xmod_layer.self_attn.k_proj.bias _lowerCamelCase = xmod_layer.self_attn.v_proj.weight _lowerCamelCase = xmod_layer.self_attn.v_proj.bias # self-attention output _lowerCamelCase = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError('''Dimensions of self-attention output weights do not match.''' ) _lowerCamelCase = xmod_layer.self_attn.out_proj.weight _lowerCamelCase = xmod_layer.self_attn.out_proj.bias _lowerCamelCase = xmod_layer.self_attn_layer_norm.weight _lowerCamelCase = xmod_layer.self_attn_layer_norm.bias # intermediate _lowerCamelCase = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('''Dimensions of intermediate weights do not match.''' ) _lowerCamelCase = xmod_layer.fca.weight _lowerCamelCase = xmod_layer.fca.bias # output _lowerCamelCase = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('''Dimensions of feed-forward weights do not match.''' ) _lowerCamelCase = xmod_layer.fca.weight _lowerCamelCase = xmod_layer.fca.bias _lowerCamelCase = xmod_layer.final_layer_norm.weight _lowerCamelCase = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: _lowerCamelCase = xmod_layer.adapter_layer_norm.weight _lowerCamelCase = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError('''Lists of language adapters do not match.''' ) for lang_code, adapter in xmod_layer.adapter_modules.items(): _lowerCamelCase = bert_output.adapter_modules[lang_code] _lowerCamelCase = xmod_layer.adapter_modules[lang_code] _lowerCamelCase = from_adapter.fca.weight _lowerCamelCase = from_adapter.fca.bias _lowerCamelCase = from_adapter.fca.weight _lowerCamelCase = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: _lowerCamelCase = xmod_sent_encoder.layer_norm.weight _lowerCamelCase = xmod_sent_encoder.layer_norm.bias if classification_head: _lowerCamelCase = xmod.model.classification_heads['''mnli'''].dense.weight _lowerCamelCase = xmod.model.classification_heads['''mnli'''].dense.bias _lowerCamelCase = xmod.model.classification_heads['''mnli'''].out_proj.weight _lowerCamelCase = xmod.model.classification_heads['''mnli'''].out_proj.bias else: # LM Head _lowerCamelCase = xmod.model.encoder.lm_head.dense.weight _lowerCamelCase = xmod.model.encoder.lm_head.dense.bias _lowerCamelCase = xmod.model.encoder.lm_head.layer_norm.weight _lowerCamelCase = xmod.model.encoder.lm_head.layer_norm.bias _lowerCamelCase = xmod.model.encoder.lm_head.weight _lowerCamelCase = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. _lowerCamelCase = xmod.encode(lowercase_ ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(lowercase_ ) _lowerCamelCase = model(lowercase_ )[0] if classification_head: _lowerCamelCase = xmod.model.classification_heads['''mnli'''](xmod.extract_features(lowercase_ ) ) else: _lowerCamelCase = xmod.model(lowercase_ , lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape , their_output.shape ) _lowerCamelCase = torch.max(torch.abs(our_output - their_output ) ).item() print(F"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 _lowerCamelCase = torch.allclose(lowercase_ , lowercase_ , atol=1e-3 ) print('''Do both models output the same tensors?''' , '''🔥''' if success else '''💩''' ) if not success: raise Exception('''Something went wRoNg''' ) Path(lowercase_ ).mkdir(parents=lowercase_ , exist_ok=lowercase_ ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--xmod_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--classification_head''', action='''store_true''', help='''Whether to convert a final classification head.''' ) __SCREAMING_SNAKE_CASE : str = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
623
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : list ) -> bool: if not isinstance(lowercase_ , lowercase_ ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(lowercase_ ) == 0: raise ValueError('''Input list must be a non empty list''' ) if len(lowercase_ ) == 1: return True _lowerCamelCase = series[1] - series[0] for index in range(len(lowercase_ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def lowerCAmelCase_( lowercase_ : list ) -> float: if not isinstance(lowercase_ , lowercase_ ): raise ValueError('''Input series is not valid, valid series - [2, 4, 6]''' ) if len(lowercase_ ) == 0: raise ValueError('''Input list must be a non empty list''' ) _lowerCamelCase = 0 for val in series: answer += val return answer / len(lowercase_ ) if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
1
"""simple docstring""" import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = GPTaTokenizer lowercase__ : Optional[int] = GPTaTokenizerFast lowercase__ : List[str] = True lowercase__ : List[Any] = {'add_prefix_space': True} lowercase__ : str = False def snake_case__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] _lowerCamelCase = {'''unk_token''': '''<unk>'''} _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(lowerCamelCase__ ) ) def snake_case__ ( self , **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def snake_case__ ( self , **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = '''lower newer''' _lowerCamelCase = '''lower newer''' return input_text, output_text def snake_case__ ( self ): _lowerCamelCase = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _lowerCamelCase = '''lower newer''' _lowerCamelCase = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokens + [tokenizer.unk_token] _lowerCamelCase = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer(add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = '''lower newer''' # Testing tokenization _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids without special tokens _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids with special tokens _lowerCamelCase = self.get_rust_tokenizer(add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_prefix_space=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing the unknown token _lowerCamelCase = tokens + [rust_tokenizer.unk_token] _lowerCamelCase = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): # It's very difficult to mix/test pretokenization with byte-level # And get both GPT2 and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def snake_case__ ( self , lowerCamelCase__=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) # Simple input _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] _lowerCamelCase = ('''This is a simple input''', '''This is a pair''') _lowerCamelCase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(lowerCamelCase__ , tokenizer_r.encode , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Simple input self.assertRaises(lowerCamelCase__ , tokenizer_r.encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Simple input self.assertRaises( lowerCamelCase__ , tokenizer_r.batch_encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' , ) # Pair input self.assertRaises(lowerCamelCase__ , tokenizer_r.encode , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Pair input self.assertRaises(lowerCamelCase__ , tokenizer_r.encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' ) # Pair input self.assertRaises( lowerCamelCase__ , tokenizer_r.batch_encode_plus , lowerCamelCase__ , max_length=lowerCamelCase__ , padding='''max_length''' , ) def snake_case__ ( self ): _lowerCamelCase = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input looooooooong''', '''This is a simple input'''] _lowerCamelCase = ('''This is a simple input''', '''This is a pair''') _lowerCamelCase = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] _lowerCamelCase = tokenizer.pad_token_id _lowerCamelCase = tokenizer(lowerCamelCase__ , padding='''max_length''' , max_length=3_0 , return_tensors='''np''' ) _lowerCamelCase = tokenizer(lowerCamelCase__ , padding=lowerCamelCase__ , truncate=lowerCamelCase__ , return_tensors='''np''' ) _lowerCamelCase = tokenizer(*lowerCamelCase__ , padding='''max_length''' , max_length=6_0 , return_tensors='''np''' ) _lowerCamelCase = tokenizer(lowerCamelCase__ , padding=lowerCamelCase__ , truncate=lowerCamelCase__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def snake_case__ ( self ): _lowerCamelCase = '''$$$''' _lowerCamelCase = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=lowerCamelCase__ , add_bos_token=lowerCamelCase__ ) _lowerCamelCase = '''This is a simple input''' _lowerCamelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = tokenizer(lowerCamelCase__ ) _lowerCamelCase = tokenizer(lowerCamelCase__ ) self.assertEqual(out_s.input_ids[0] , lowerCamelCase__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _lowerCamelCase = tokenizer.decode(out_s.input_ids ) _lowerCamelCase = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , lowerCamelCase__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def snake_case__ ( self ): pass def snake_case__ ( self ): # TODO: change to self.get_tokenizers() when the fast version is implemented _lowerCamelCase = [self.get_tokenizer(do_lower_case=lowerCamelCase__ , add_bos_token=lowerCamelCase__ )] for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): _lowerCamelCase = '''Encode this.''' _lowerCamelCase = '''This one too please.''' _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) encoded_sequence += tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode_plus( lowerCamelCase__ , lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_special_tokens_mask=lowerCamelCase__ , ) _lowerCamelCase = encoded_sequence_dict['''input_ids'''] _lowerCamelCase = encoded_sequence_dict['''special_tokens_mask'''] self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) _lowerCamelCase = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(lowerCamelCase__ ) ] _lowerCamelCase = [x for x in filtered_sequence if x is not None] self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @require_tokenizers class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # More context: # https://huggingface.co/wjmcat/opt-350m-paddle/discussions/1 # https://huggingface.slack.com/archives/C01N44FJDHT/p1653511495183519 # https://github.com/huggingface/transformers/pull/17088#discussion_r871246439 _lowerCamelCase = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=lowerCamelCase__ ) _lowerCamelCase = '''A photo of a cat''' _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) self.assertEqual(lowerCamelCase__ , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('''test_opt''' ) _lowerCamelCase = AutoTokenizer.from_pretrained('''./test_opt''' ) _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) self.assertEqual(lowerCamelCase__ , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) def snake_case__ ( self ): _lowerCamelCase = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , use_slow=lowerCamelCase__ ) _lowerCamelCase = '''A photo of a cat''' _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) # Same as above self.assertEqual(lowerCamelCase__ , [2, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) @unittest.skip('''This test is failing because of a bug in the fast tokenizer''' ) def snake_case__ ( self ): _lowerCamelCase = AutoTokenizer.from_pretrained('''facebook/opt-350m''' , from_slow=lowerCamelCase__ ) _lowerCamelCase = '''bos''' _lowerCamelCase = tokenizer.get_vocab()['''bos'''] _lowerCamelCase = '''A photo of a cat''' _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) # We changed the bos token self.assertEqual(lowerCamelCase__ , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] ) tokenizer.save_pretrained('''./tok''' ) _lowerCamelCase = AutoTokenizer.from_pretrained('''./tok''' ) self.assertTrue(tokenizer.is_fast ) _lowerCamelCase = tokenizer.encode( lowerCamelCase__ , ) self.assertEqual(lowerCamelCase__ , [3_1_9_5_7, 2_5_0, 1_3_4_5, 9, 1_0, 4_7_5_8] )
623
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
623
1
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = '''ylacombe/bark-small''' _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = '''en_speaker_1''' _lowerCamelCase = '''This is a test string''' _lowerCamelCase = '''speaker_embeddings_path.json''' _lowerCamelCase = '''speaker_embeddings''' def snake_case__ ( self , **lowerCamelCase__ ): return AutoTokenizer.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def snake_case__ ( self ): shutil.rmtree(self.tmpdirname ) def snake_case__ ( self ): _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = BarkProcessor(tokenizer=lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def snake_case__ ( self ): _lowerCamelCase = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) _lowerCamelCase = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) _lowerCamelCase = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token='''(BOS)''' , eos_token='''(EOS)''' , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def snake_case__ ( self ): _lowerCamelCase = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) _lowerCamelCase = 3_5 _lowerCamelCase = 2 _lowerCamelCase = 8 _lowerCamelCase = { '''semantic_prompt''': np.ones(lowerCamelCase__ ), '''coarse_prompt''': np.ones((nb_codebooks_coarse, seq_len) ), '''fine_prompt''': np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset _lowerCamelCase = processor(text=self.input_string , voice_preset=lowerCamelCase__ ) _lowerCamelCase = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCamelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from npz file _lowerCamelCase = os.path.join(self.tmpdirname , '''file.npz''' ) np.savez(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = processor(text=self.input_string , voice_preset=lowerCamelCase__ ) _lowerCamelCase = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCamelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from the hub _lowerCamelCase = processor(text=self.input_string , voice_preset=self.voice_preset ) def snake_case__ ( self ): _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = BarkProcessor(tokenizer=lowerCamelCase__ ) _lowerCamelCase = processor(text=self.input_string ) _lowerCamelCase = tokenizer( self.input_string , padding='''max_length''' , max_length=2_5_6 , add_special_tokens=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
623
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int ) -> list[int]: if num <= 0: raise ValueError('''Input must be a positive integer''' ) _lowerCamelCase = [True] * (num + 1) _lowerCamelCase = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , lowercase_ ): _lowerCamelCase = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : Dict = int(input('''Enter a positive integer: ''').strip()) print(prime_sieve_eratosthenes(user_num))
623
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
1
"""simple docstring""" import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer __SCREAMING_SNAKE_CASE : int = '''bart''' __SCREAMING_SNAKE_CASE : int = True @st.cache(allow_output_mutation=lowercase_ ) def lowerCAmelCase_( ) -> Tuple: if LOAD_DENSE_INDEX: _lowerCamelCase = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) _lowerCamelCase = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) _lowerCamelCase = qar_model.eval() else: _lowerCamelCase , _lowerCamelCase = (None, None) if MODEL_TYPE == "bart": _lowerCamelCase = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) _lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) _lowerCamelCase = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) _lowerCamelCase = sas_model.eval() else: _lowerCamelCase , _lowerCamelCase = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=lowercase_ ) def lowerCAmelCase_( ) -> Union[str, Any]: if LOAD_DENSE_INDEX: _lowerCamelCase = faiss.StandardGpuResources() _lowerCamelCase = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] _lowerCamelCase = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 1_28) , ) _lowerCamelCase = faiss.IndexFlatIP(1_28 ) _lowerCamelCase = faiss.index_cpu_to_gpu(lowercase_ , 1 , lowercase_ ) wikiaab_gpu_index_flat.add(lowercase_ ) # TODO fix for larger GPU else: _lowerCamelCase , _lowerCamelCase = (None, None) _lowerCamelCase = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=lowercase_ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) _lowerCamelCase = elia['''train_eli5'''] _lowerCamelCase = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 1_28) ) _lowerCamelCase = faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(lowercase_ ) return (elia_train, eli5_train_q_index) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = load_indexes() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = load_models() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : str = load_train_data() def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Any=10 ) -> int: _lowerCamelCase = embed_questions_for_retrieval([question] , lowercase_ , lowercase_ ) _lowerCamelCase , _lowerCamelCase = eli5_train_q_index.search(lowercase_ , lowercase_ ) _lowerCamelCase = [elia_train[int(lowercase_ )] for i in I[0]] return nn_examples def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Optional[Any]="wiki40b" , lowercase_ : Dict="dense" , lowercase_ : Dict=10 ) -> int: if source == "none": _lowerCamelCase , _lowerCamelCase = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": _lowerCamelCase , _lowerCamelCase = query_qa_dense_index( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: _lowerCamelCase , _lowerCamelCase = query_es_index( lowercase_ , lowercase_ , index_name='''english_wiki40b_snippets_100w''' , n_results=lowercase_ , ) _lowerCamelCase = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] _lowerCamelCase = '''question: {} context: {}'''.format(lowercase_ , lowercase_ ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda lowercase_ : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda lowercase_ : None), } ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Any=64 , lowercase_ : Union[str, Any]=2_56 , lowercase_ : List[str]=False , lowercase_ : Optional[int]=2 , lowercase_ : Union[str, Any]=0.9_5 , lowercase_ : List[str]=0.8 ) -> Dict: with torch.no_grad(): _lowerCamelCase = qa_sas_generate( lowercase_ , lowercase_ , lowercase_ , num_answers=1 , num_beams=lowercase_ , min_len=lowercase_ , max_len=lowercase_ , do_sample=lowercase_ , temp=lowercase_ , top_p=lowercase_ , top_k=lowercase_ , max_input_length=10_24 , device='''cuda:0''' , )[0] return (answer, support_list) st.title('''Long Form Question Answering with ELI5''') # Start sidebar __SCREAMING_SNAKE_CASE : Any = '''<img src=\'https://huggingface.co/front/assets/huggingface_logo.svg\'>''' __SCREAMING_SNAKE_CASE : int = ''' <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class="img-container"> <!-- Inline parent element --> %s </span> </body> </html> ''' % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia __SCREAMING_SNAKE_CASE : int = ''' This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. ''' st.sidebar.markdown(description, unsafe_allow_html=True) __SCREAMING_SNAKE_CASE : List[str] = [ '''Answer the question''', '''View the retrieved document only''', '''View the most similar ELI5 question and answer''', '''Show me everything, please!''', ] __SCREAMING_SNAKE_CASE : Optional[int] = st.sidebar.checkbox('''Demo options''') if demo_options: __SCREAMING_SNAKE_CASE : Dict = st.sidebar.selectbox( '''''', action_list, index=3, ) __SCREAMING_SNAKE_CASE : Any = action_list.index(action_st) __SCREAMING_SNAKE_CASE : str = st.sidebar.selectbox( '''''', ['''Show full text of passages''', '''Show passage section titles'''], index=0, ) __SCREAMING_SNAKE_CASE : Optional[int] = show_type == '''Show full text of passages''' else: __SCREAMING_SNAKE_CASE : Any = 3 __SCREAMING_SNAKE_CASE : Union[str, Any] = True __SCREAMING_SNAKE_CASE : Tuple = st.sidebar.checkbox('''Retrieval options''') if retrieval_options: __SCREAMING_SNAKE_CASE : Optional[int] = ''' ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. ''' st.sidebar.markdown(retriever_info) __SCREAMING_SNAKE_CASE : Any = st.sidebar.selectbox('''Which Wikipedia format should the model use?''', ['''wiki40b''', '''none''']) __SCREAMING_SNAKE_CASE : Optional[Any] = st.sidebar.selectbox('''Which Wikipedia indexer should the model use?''', ['''dense''', '''sparse''', '''mixed''']) else: __SCREAMING_SNAKE_CASE : Optional[Any] = '''wiki40b''' __SCREAMING_SNAKE_CASE : Dict = '''dense''' __SCREAMING_SNAKE_CASE : Any = '''beam''' __SCREAMING_SNAKE_CASE : int = 2 __SCREAMING_SNAKE_CASE : Dict = 6_4 __SCREAMING_SNAKE_CASE : int = 2_5_6 __SCREAMING_SNAKE_CASE : Tuple = None __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : Tuple = st.sidebar.checkbox('''Generation options''') if generate_options: __SCREAMING_SNAKE_CASE : int = ''' ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder\'s output probabilities. ''' st.sidebar.markdown(generate_info) __SCREAMING_SNAKE_CASE : Union[str, Any] = st.sidebar.selectbox('''Would you like to use beam search or sample an answer?''', ['''beam''', '''sampled''']) __SCREAMING_SNAKE_CASE : Optional[Any] = st.sidebar.slider( '''Minimum generation length''', min_value=8, max_value=2_5_6, value=6_4, step=8, format=None, key=None ) __SCREAMING_SNAKE_CASE : Optional[int] = st.sidebar.slider( '''Maximum generation length''', min_value=6_4, max_value=5_1_2, value=2_5_6, step=1_6, format=None, key=None ) if sampled == "beam": __SCREAMING_SNAKE_CASE : str = st.sidebar.slider('''Beam size''', min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: __SCREAMING_SNAKE_CASE : str = st.sidebar.slider( '''Nucleus sampling p''', min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) __SCREAMING_SNAKE_CASE : Optional[Any] = st.sidebar.slider( '''Temperature''', min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) __SCREAMING_SNAKE_CASE : Union[str, Any] = None # start main text __SCREAMING_SNAKE_CASE : Optional[Any] = [ '''<MY QUESTION>''', '''How do people make chocolate?''', '''Why do we get a fever when we are sick?''', '''How can different animals perceive different colors?''', '''What is natural language processing?''', '''What\'s the best way to treat a sunburn?''', '''What exactly are vitamins ?''', '''How does nuclear energy provide electricity?''', '''What\'s the difference between viruses and bacteria?''', '''Why are flutes classified as woodwinds when most of them are made out of metal ?''', '''Why do people like drinking coffee even though it tastes so bad?''', '''What happens when wine ages? How does it make the wine taste better?''', '''If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?''', '''How can we set a date to the beginning or end of an artistic period? Doesn\'t the change happen gradually?''', '''How does New Zealand have so many large bird predators?''', ] __SCREAMING_SNAKE_CASE : int = st.selectbox( '''What would you like to ask? ---- select <MY QUESTION> to enter a new query''', questions_list, index=1, ) if question_s == "<MY QUESTION>": __SCREAMING_SNAKE_CASE : Optional[int] = st.text_input('''Enter your question here:''', '''''') else: __SCREAMING_SNAKE_CASE : Optional[int] = question_s if st.button('''Show me!'''): if action in [0, 1, 3]: if index_type == "mixed": __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = make_support(question, source=wiki_source, method='''dense''', n_results=1_0) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = make_support(question, source=wiki_source, method='''sparse''', n_results=1_0) __SCREAMING_SNAKE_CASE : List[str] = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] __SCREAMING_SNAKE_CASE : Any = support_list[:1_0] __SCREAMING_SNAKE_CASE : Optional[int] = '''<P> ''' + ''' <P> '''.join([res[-1] for res in support_list]) else: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[Any] = make_support(question, source=wiki_source, method=index_type, n_results=1_0) if action in [0, 3]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == '''sampled'''), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown('''### The model generated answer is:''') st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown('''--- \n ### The model is drawing information from the following Wikipedia passages:''') for i, res in enumerate(support_list): __SCREAMING_SNAKE_CASE : int = '''https://en.wikipedia.org/wiki/{}'''.format(res[0].replace(''' ''', '''_''')) __SCREAMING_SNAKE_CASE : Any = res[1].strip() if sec_titles == "": __SCREAMING_SNAKE_CASE : Dict = '''[{}]({})'''.format(res[0], wiki_url) else: __SCREAMING_SNAKE_CASE : List[str] = sec_titles.split(''' & ''') __SCREAMING_SNAKE_CASE : List[str] = ''' & '''.join( ['''[{}]({}#{})'''.format(sec.strip(), wiki_url, sec.strip().replace(''' ''', '''_''')) for sec in sec_list] ) st.markdown( '''{0:02d} - **Article**: {1:<18} <br> _Section_: {2}'''.format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( '''> <span style="font-family:arial; font-size:10pt;">''' + res[-1] + '''</span>''', unsafe_allow_html=True ) if action in [2, 3]: __SCREAMING_SNAKE_CASE : Dict = find_nearest_training(question) __SCREAMING_SNAKE_CASE : Optional[Any] = nn_train_list[0] st.markdown( '''--- \n ### The most similar question in the ELI5 training set was: \n\n {}'''.format(train_exple['''title''']) ) __SCREAMING_SNAKE_CASE : Optional[int] = [ '''{}. {}'''.format(i + 1, ''' \n'''.join([line.strip() for line in ans.split('''\n''') if line.strip() != ''''''])) for i, (ans, sc) in enumerate(zip(train_exple['''answers''']['''text'''], train_exple['''answers''']['''score'''])) if i == 0 or sc > 2 ] st.markdown('''##### Its answers were: \n\n {}'''.format('''\n'''.join(answers_st))) __SCREAMING_SNAKE_CASE : str = ''' --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* ''' st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str ) -> str: return " ".join( ''''''.join(word[::-1] ) if len(lowercase_ ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class lowerCamelCase_( tf.keras.optimizers.schedules.LearningRateSchedule ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): super().__init__() _lowerCamelCase = initial_learning_rate _lowerCamelCase = warmup_steps _lowerCamelCase = power _lowerCamelCase = decay_schedule_fn _lowerCamelCase = name def __call__( self , lowerCamelCase__ ): with tf.name_scope(self.name or '''WarmUp''' ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. _lowerCamelCase = tf.cast(lowerCamelCase__ , tf.floataa ) _lowerCamelCase = tf.cast(self.warmup_steps , tf.floataa ) _lowerCamelCase = global_step_float / warmup_steps_float _lowerCamelCase = self.initial_learning_rate * tf.math.pow(lowerCamelCase__ , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=lowerCamelCase__ , ) def snake_case__ ( self ): return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def lowerCAmelCase_( lowercase_ : float , lowercase_ : int , lowercase_ : int , lowercase_ : float = 0.0 , lowercase_ : float = 0.9 , lowercase_ : float = 0.9_9_9 , lowercase_ : float = 1e-8 , lowercase_ : Optional[float] = None , lowercase_ : Optional[float] = None , lowercase_ : float = 0.0 , lowercase_ : float = 1.0 , lowercase_ : Optional[List[str]] = None , ) -> List[Any]: _lowerCamelCase = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=lowercase_ , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=lowercase_ , ) if num_warmup_steps: _lowerCamelCase = WarmUp( initial_learning_rate=lowercase_ , decay_schedule_fn=lowercase_ , warmup_steps=lowercase_ , ) if weight_decay_rate > 0.0: _lowerCamelCase = AdamWeightDecay( learning_rate=lowercase_ , weight_decay_rate=lowercase_ , beta_a=lowercase_ , beta_a=lowercase_ , epsilon=lowercase_ , clipnorm=lowercase_ , global_clipnorm=lowercase_ , exclude_from_weight_decay=['''LayerNorm''', '''layer_norm''', '''bias'''] , include_in_weight_decay=lowercase_ , ) else: _lowerCamelCase = tf.keras.optimizers.Adam( learning_rate=lowercase_ , beta_a=lowercase_ , beta_a=lowercase_ , epsilon=lowercase_ , clipnorm=lowercase_ , global_clipnorm=lowercase_ , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ = 0.0_0_1 , lowerCamelCase__ = 0.9 , lowerCamelCase__ = 0.9_9_9 , lowerCamelCase__ = 1e-7 , lowerCamelCase__ = False , lowerCamelCase__ = 0.0 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = "AdamWeightDecay" , **lowerCamelCase__ , ): super().__init__(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = weight_decay_rate _lowerCamelCase = include_in_weight_decay _lowerCamelCase = exclude_from_weight_decay @classmethod def snake_case__ ( cls , lowerCamelCase__ ): _lowerCamelCase = {'''WarmUp''': WarmUp} return super(lowerCamelCase__ , cls ).from_config(lowerCamelCase__ , custom_objects=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): super(lowerCamelCase__ , self )._prepare_local(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tf.constant( self.weight_decay_rate , name='''adam_weight_decay_rate''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]['''weight_decay_rate'''] , use_locking=self._use_locking , ) return tf.no_op() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): _lowerCamelCase , _lowerCamelCase = list(zip(*lowerCamelCase__ ) ) return super(lowerCamelCase__ , self ).apply_gradients(zip(lowerCamelCase__ , lowerCamelCase__ ) , name=lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if apply_state is None: return self._decayed_lr_t[var_dtype], {} _lowerCamelCase = apply_state or {} _lowerCamelCase = apply_state.get((var_device, var_dtype) ) if coefficients is None: _lowerCamelCase = self._fallback_apply_state(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase , _lowerCamelCase = self._get_lr(var.device , var.dtype.base_dtype , lowerCamelCase__ ) _lowerCamelCase = self._decay_weights_op(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) with tf.control_dependencies([decay] ): return super(lowerCamelCase__ , self )._resource_apply_dense(lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase , _lowerCamelCase = self._get_lr(var.device , var.dtype.base_dtype , lowerCamelCase__ ) _lowerCamelCase = self._decay_weights_op(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) with tf.control_dependencies([decay] ): return super(lowerCamelCase__ , self )._resource_apply_sparse(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = super().get_config() config.update({'''weight_decay_rate''': self.weight_decay_rate} ) return config def snake_case__ ( self , lowerCamelCase__ ): if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(lowerCamelCase__ , lowerCamelCase__ ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(lowerCamelCase__ , lowerCamelCase__ ) is not None: return False return True class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self ): _lowerCamelCase = [] _lowerCamelCase = None @property def snake_case__ ( self ): if self._accum_steps is None: _lowerCamelCase = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=lowerCamelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def snake_case__ ( self ): if not self._gradients: raise ValueError('''The accumulator should be called first to initialize the gradients''' ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self , lowerCamelCase__ ): if not self._gradients: _lowerCamelCase = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(lowerCamelCase__ ) , trainable=lowerCamelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(lowerCamelCase__ ) != len(self._gradients ): raise ValueError(F"""Expected {len(self._gradients )} gradients, but got {len(lowerCamelCase__ )}""" ) for accum_gradient, gradient in zip(self._gradients , lowerCamelCase__ ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(lowerCamelCase__ ) self._accum_steps.assign_add(1 ) def snake_case__ ( self ): if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(lowerCamelCase__ ) )
623
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
1
"""simple docstring""" import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Optional[Any]=None ) -> Optional[Any]: # set parameter of one layer assert torch_layer.weight.shape == weight.shape, F"""{torch_layer} layer.weight does not match""" _lowerCamelCase = nn.Parameter(lowercase_ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F"""{torch_layer} layer.bias does not match""" _lowerCamelCase = nn.Parameter(lowercase_ ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : str ) -> Any: # set torch weights for 1-to-1 comparison _lowerCamelCase = np.asarray(weights[0] ) _lowerCamelCase = np.asarray(weights[1] ) _lowerCamelCase = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(lowercase_ ).transpose(1 , 2 ).contiguous().view(-1 , lowercase_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(lowercase_ ).transpose(1 , 2 ).contiguous().view(-1 , lowercase_ ) , ) set_param( torch_layer.output.dense , torch.tensor(lowercase_ ).view(-1 , lowercase_ ).contiguous().transpose(0 , 1 ) , ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : Dict ) -> Optional[int]: # set torch weights for 1-to-1 comparison _lowerCamelCase = np.asarray(weights[0] ) _lowerCamelCase = np.asarray(weights[1] ) _lowerCamelCase = np.asarray(weights[2] ) _lowerCamelCase = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(lowercase_ ).transpose(1 , 2 ).contiguous().view(-1 , lowercase_ ) , ) set_param( torch_layer.self_attention.key , torch.tensor(lowercase_ ).transpose(1 , 2 ).contiguous().view(-1 , lowercase_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(lowercase_ ).transpose(1 , 2 ).contiguous().view(-1 , lowercase_ ) , ) set_param( torch_layer.output.dense , torch.tensor(lowercase_ ).view(-1 , lowercase_ ).contiguous().transpose(0 , 1 ) , ) def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : int ) -> int: # layernorm 1 _lowerCamelCase = weights[0][0][0] _lowerCamelCase = np.asarray(layer_norm_a[0] ) _lowerCamelCase = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(lowercase_ ) , torch.tensor(lowercase_ ) , ) # lsh weights + output _lowerCamelCase = weights[0][1] if len(lowercase_ ) < 4: set_layer_weights_in_torch_lsh(lowercase_ , torch_block.attention , lowercase_ ) else: set_layer_weights_in_torch_local(lowercase_ , torch_block.attention , lowercase_ ) # intermediate weighs _lowerCamelCase = weights[2][0][1][2] # Chunked Feed Forward if len(lowercase_ ) == 4: _lowerCamelCase = intermediate_weights[2] # layernorm 2 _lowerCamelCase = np.asarray(intermediate_weights[0][0] ) _lowerCamelCase = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(lowercase_ ) , torch.tensor(lowercase_ ) , ) # intermediate dense _lowerCamelCase = np.asarray(intermediate_weights[1][0] ) _lowerCamelCase = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(lowercase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(lowercase_ ) , ) # intermediate out _lowerCamelCase = np.asarray(intermediate_weights[4][0] ) _lowerCamelCase = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(lowercase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(lowercase_ ) , ) def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Any ) -> str: # reformer model _lowerCamelCase = torch_model.reformer # word embeds _lowerCamelCase = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(lowercase_ ) , ) if isinstance(weights[3] , lowercase_ ): _lowerCamelCase = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _lowerCamelCase = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F"""{position_embeddings[emb_idx]} emb does not match""" _lowerCamelCase = nn.Parameter(torch.tensor(lowercase_ ) ) _lowerCamelCase = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( lowercase_ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _lowerCamelCase = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(lowercase_ , lowercase_ , lowercase_ ) # output layer norm _lowerCamelCase = np.asarray(weights[7][0] ) _lowerCamelCase = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(lowercase_ ) , torch.tensor(lowercase_ ) , ) # output embeddings _lowerCamelCase = np.asarray(weights[9][0] ) _lowerCamelCase = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(lowercase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(lowercase_ ) , ) def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Union[str, Any] ) -> Optional[Any]: # Initialise PyTorch model _lowerCamelCase = ReformerConfig.from_json_file(lowercase_ ) print(F"""Building PyTorch model from configuration: {config}""" ) _lowerCamelCase = ReformerModelWithLMHead(lowercase_ ) with open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = pickle.load(lowercase_ )['''weights'''] set_model_weights_in_torch(lowercase_ , lowercase_ , config.hidden_size ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--trax_model_pkl_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained Reformer model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
623
"""simple docstring""" from typing import Any class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) _lowerCamelCase = temp.next print() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = Node(lowerCamelCase__ ) _lowerCamelCase = self.head _lowerCamelCase = new_node def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if node_data_a == node_data_a: return else: _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next if node_a is None or node_a is None: return _lowerCamelCase , _lowerCamelCase = node_a.data, node_a.data if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
623
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = '''▁''' __SCREAMING_SNAKE_CASE : Union[str, Any] = {'''vocab_file''': '''sentencepiece.bpe.model'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model''', } } __SCREAMING_SNAKE_CASE : str = { '''facebook/xglm-564M''': 2_0_4_8, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Any = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__ = None , **lowerCamelCase__ , ): _lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer _lowerCamelCase = 7 _lowerCamelCase = [F"""<madeupword{i}>""" for i in range(self.num_madeup_words )] _lowerCamelCase = kwargs.get('''additional_special_tokens''' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase__ ) ) _lowerCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _lowerCamelCase = 1 # Mimic fairseq token-to-id alignment for the first 4 token _lowerCamelCase = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} _lowerCamelCase = len(self.sp_model ) _lowerCamelCase = {F"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): _lowerCamelCase = self.__dict__.copy() _lowerCamelCase = None _lowerCamelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self , lowerCamelCase__ ): _lowerCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _lowerCamelCase = {} _lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if token_ids_a is None: return [self.sep_token_id] + token_ids_a _lowerCamelCase = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def snake_case__ ( self ): return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def snake_case__ ( self ): _lowerCamelCase = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def snake_case__ ( self , lowerCamelCase__ ): return self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowerCamelCase = self.sp_model.PieceToId(lowerCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def snake_case__ ( self , lowerCamelCase__ ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ).replace(lowerCamelCase__ , ''' ''' ).strip() return out_string def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , '''wb''' ) as fi: _lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,)
623
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __SCREAMING_SNAKE_CASE : Optional[Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[Any]: from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> List[str]: from diffusers.utils.testing_utils import pytest_terminal_summary_main _lowerCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
623
1
"""simple docstring""" from scipy.stats import spearmanr import datasets __SCREAMING_SNAKE_CASE : Optional[Any] = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __SCREAMING_SNAKE_CASE : Tuple = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __SCREAMING_SNAKE_CASE : Dict = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = spearmanr(lowerCamelCase__ , lowerCamelCase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
623
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
1
"""simple docstring""" from typing import Any def lowerCAmelCase_( lowercase_ : list ) -> list[Any]: if not input_list: return [] _lowerCamelCase = [input_list.count(lowercase_ ) for value in input_list] _lowerCamelCase = max(lowercase_ ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(lowercase_ ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
1
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
1
"""simple docstring""" import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : List[str]=1 ) -> List[Any]: if n_shave_prefix_segments >= 0: return ".".join(path.split('''.''' )[n_shave_prefix_segments:] ) else: return ".".join(path.split('''.''' )[:n_shave_prefix_segments] ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Any=0 ) -> Tuple: _lowerCamelCase = [] for old_item in old_list: _lowerCamelCase = old_item.replace('''in_layers.0''' , '''norm1''' ) _lowerCamelCase = new_item.replace('''in_layers.2''' , '''conv1''' ) _lowerCamelCase = new_item.replace('''out_layers.0''' , '''norm2''' ) _lowerCamelCase = new_item.replace('''out_layers.3''' , '''conv2''' ) _lowerCamelCase = new_item.replace('''emb_layers.1''' , '''time_emb_proj''' ) _lowerCamelCase = new_item.replace('''skip_connection''' , '''conv_shortcut''' ) _lowerCamelCase = shave_segments(lowercase_ , n_shave_prefix_segments=lowercase_ ) mapping.append({'''old''': old_item, '''new''': new_item} ) return mapping def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : List[Any]=0 ) -> List[Any]: _lowerCamelCase = [] for old_item in old_list: _lowerCamelCase = old_item _lowerCamelCase = new_item.replace('''norm.weight''' , '''group_norm.weight''' ) _lowerCamelCase = new_item.replace('''norm.bias''' , '''group_norm.bias''' ) _lowerCamelCase = new_item.replace('''proj_out.weight''' , '''proj_attn.weight''' ) _lowerCamelCase = new_item.replace('''proj_out.bias''' , '''proj_attn.bias''' ) _lowerCamelCase = shave_segments(lowercase_ , n_shave_prefix_segments=lowercase_ ) mapping.append({'''old''': old_item, '''new''': new_item} ) return mapping def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Union[str, Any] , lowercase_ : int , lowercase_ : str=None , lowercase_ : Tuple=None , lowercase_ : Optional[Any]=None ) -> Optional[Any]: assert isinstance(lowercase_ , lowercase_ ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): _lowerCamelCase = old_checkpoint[path] _lowerCamelCase = old_tensor.shape[0] // 3 _lowerCamelCase = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) _lowerCamelCase = old_tensor.shape[0] // config['''num_head_channels'''] // 3 _lowerCamelCase = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = old_tensor.split(channels // num_heads , dim=1 ) _lowerCamelCase = query.reshape(lowercase_ ) _lowerCamelCase = key.reshape(lowercase_ ) _lowerCamelCase = value.reshape(lowercase_ ) for path in paths: _lowerCamelCase = path['''new'''] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here _lowerCamelCase = new_path.replace('''middle_block.0''' , '''mid_block.resnets.0''' ) _lowerCamelCase = new_path.replace('''middle_block.1''' , '''mid_block.attentions.0''' ) _lowerCamelCase = new_path.replace('''middle_block.2''' , '''mid_block.resnets.1''' ) if additional_replacements is not None: for replacement in additional_replacements: _lowerCamelCase = new_path.replace(replacement['''old'''] , replacement['''new'''] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: _lowerCamelCase = old_checkpoint[path['''old''']][:, :, 0] else: _lowerCamelCase = old_checkpoint[path['''old''']] def lowerCAmelCase_( lowercase_ : int , lowercase_ : Tuple ) -> Dict: _lowerCamelCase = {} _lowerCamelCase = checkpoint['''time_embed.0.weight'''] _lowerCamelCase = checkpoint['''time_embed.0.bias'''] _lowerCamelCase = checkpoint['''time_embed.2.weight'''] _lowerCamelCase = checkpoint['''time_embed.2.bias'''] _lowerCamelCase = checkpoint['''input_blocks.0.0.weight'''] _lowerCamelCase = checkpoint['''input_blocks.0.0.bias'''] _lowerCamelCase = checkpoint['''out.0.weight'''] _lowerCamelCase = checkpoint['''out.0.bias'''] _lowerCamelCase = checkpoint['''out.2.weight'''] _lowerCamelCase = checkpoint['''out.2.bias'''] # Retrieves the keys for the input blocks only _lowerCamelCase = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''input_blocks''' in layer} ) _lowerCamelCase = { layer_id: [key for key in checkpoint if F"""input_blocks.{layer_id}""" in key] for layer_id in range(lowercase_ ) } # Retrieves the keys for the middle blocks only _lowerCamelCase = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''middle_block''' in layer} ) _lowerCamelCase = { layer_id: [key for key in checkpoint if F"""middle_block.{layer_id}""" in key] for layer_id in range(lowercase_ ) } # Retrieves the keys for the output blocks only _lowerCamelCase = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''output_blocks''' in layer} ) _lowerCamelCase = { layer_id: [key for key in checkpoint if F"""output_blocks.{layer_id}""" in key] for layer_id in range(lowercase_ ) } for i in range(1 , lowercase_ ): _lowerCamelCase = (i - 1) // (config['''num_res_blocks'''] + 1) _lowerCamelCase = (i - 1) % (config['''num_res_blocks'''] + 1) _lowerCamelCase = [key for key in input_blocks[i] if F"""input_blocks.{i}.0""" in key] _lowerCamelCase = [key for key in input_blocks[i] if F"""input_blocks.{i}.1""" in key] if F"""input_blocks.{i}.0.op.weight""" in checkpoint: _lowerCamelCase = checkpoint[ F"""input_blocks.{i}.0.op.weight""" ] _lowerCamelCase = checkpoint[ F"""input_blocks.{i}.0.op.bias""" ] continue _lowerCamelCase = renew_resnet_paths(lowercase_ ) _lowerCamelCase = {'''old''': F"""input_blocks.{i}.0""", '''new''': F"""down_blocks.{block_id}.resnets.{layer_in_block_id}"""} _lowerCamelCase = {'''old''': '''resnets.2.op''', '''new''': '''downsamplers.0.op'''} assign_to_checkpoint( lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path, resnet_op] , config=lowercase_ ) if len(lowercase_ ): _lowerCamelCase = renew_attention_paths(lowercase_ ) _lowerCamelCase = { '''old''': F"""input_blocks.{i}.1""", '''new''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}""", } _lowerCamelCase = { F"""input_blocks.{i}.1.qkv.bias""": { '''key''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", '''query''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", '''value''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""input_blocks.{i}.1.qkv.weight""": { '''key''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", '''query''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", '''value''': F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , attention_paths_to_split=lowercase_ , config=lowercase_ , ) _lowerCamelCase = middle_blocks[0] _lowerCamelCase = middle_blocks[1] _lowerCamelCase = middle_blocks[2] _lowerCamelCase = renew_resnet_paths(lowercase_ ) assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , config=lowercase_ ) _lowerCamelCase = renew_resnet_paths(lowercase_ ) assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , config=lowercase_ ) _lowerCamelCase = renew_attention_paths(lowercase_ ) _lowerCamelCase = { '''middle_block.1.qkv.bias''': { '''key''': '''mid_block.attentions.0.key.bias''', '''query''': '''mid_block.attentions.0.query.bias''', '''value''': '''mid_block.attentions.0.value.bias''', }, '''middle_block.1.qkv.weight''': { '''key''': '''mid_block.attentions.0.key.weight''', '''query''': '''mid_block.attentions.0.query.weight''', '''value''': '''mid_block.attentions.0.value.weight''', }, } assign_to_checkpoint( lowercase_ , lowercase_ , lowercase_ , attention_paths_to_split=lowercase_ , config=lowercase_ ) for i in range(lowercase_ ): _lowerCamelCase = i // (config['''num_res_blocks'''] + 1) _lowerCamelCase = i % (config['''num_res_blocks'''] + 1) _lowerCamelCase = [shave_segments(lowercase_ , 2 ) for name in output_blocks[i]] _lowerCamelCase = {} for layer in output_block_layers: _lowerCamelCase , _lowerCamelCase = layer.split('''.''' )[0], shave_segments(lowercase_ , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(lowercase_ ) else: _lowerCamelCase = [layer_name] if len(lowercase_ ) > 1: _lowerCamelCase = [key for key in output_blocks[i] if F"""output_blocks.{i}.0""" in key] _lowerCamelCase = [key for key in output_blocks[i] if F"""output_blocks.{i}.1""" in key] _lowerCamelCase = renew_resnet_paths(lowercase_ ) _lowerCamelCase = renew_resnet_paths(lowercase_ ) _lowerCamelCase = {'''old''': F"""output_blocks.{i}.0""", '''new''': F"""up_blocks.{block_id}.resnets.{layer_in_block_id}"""} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) if ["conv.weight", "conv.bias"] in output_block_list.values(): _lowerCamelCase = list(output_block_list.values() ).index(['''conv.weight''', '''conv.bias'''] ) _lowerCamelCase = checkpoint[ F"""output_blocks.{i}.{index}.conv.weight""" ] _lowerCamelCase = checkpoint[ F"""output_blocks.{i}.{index}.conv.bias""" ] # Clear attentions as they have been attributed above. if len(lowercase_ ) == 2: _lowerCamelCase = [] if len(lowercase_ ): _lowerCamelCase = renew_attention_paths(lowercase_ ) _lowerCamelCase = { '''old''': F"""output_blocks.{i}.1""", '''new''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}""", } _lowerCamelCase = { F"""output_blocks.{i}.1.qkv.bias""": { '''key''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", '''query''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", '''value''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""output_blocks.{i}.1.qkv.weight""": { '''key''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", '''query''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", '''value''': F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any('''qkv''' in key for key in attentions ) else None , config=lowercase_ , ) else: _lowerCamelCase = renew_resnet_paths(lowercase_ , n_shave_prefix_segments=1 ) for path in resnet_0_paths: _lowerCamelCase = '''.'''.join(['''output_blocks''', str(lowercase_ ), path['''old''']] ) _lowerCamelCase = '''.'''.join(['''up_blocks''', str(lowercase_ ), '''resnets''', str(lowercase_ ), path['''new''']] ) _lowerCamelCase = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') __SCREAMING_SNAKE_CASE : List[str] = parser.parse_args() __SCREAMING_SNAKE_CASE : List[Any] = torch.load(args.checkpoint_path) with open(args.config_file) as f: __SCREAMING_SNAKE_CASE : Optional[Any] = json.loads(f.read()) __SCREAMING_SNAKE_CASE : int = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] __SCREAMING_SNAKE_CASE : List[str] = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: __SCREAMING_SNAKE_CASE : int = DDPMScheduler.from_config('''/'''.join(args.checkpoint_path.split('''/''')[:-1])) __SCREAMING_SNAKE_CASE : Any = VQModel.from_pretrained('''/'''.join(args.checkpoint_path.split('''/''')[:-1])) __SCREAMING_SNAKE_CASE : List[str] = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
623
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
1
"""simple docstring""" import itertools import math def lowerCAmelCase_( lowercase_ : int ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowercase_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCAmelCase_( ) -> List[str]: _lowerCamelCase = 2 while True: if is_prime(lowercase_ ): yield num num += 1 def lowerCAmelCase_( lowercase_ : int = 1_00_01 ) -> int: return next(itertools.islice(prime_generator() , nth - 1 , lowercase_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
623
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : torch.FloatTensor class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__=3 , lowerCamelCase__=3 , lowerCamelCase__=("DownEncoderBlock2D",) , lowerCamelCase__=(6_4,) , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__="silu" , lowerCamelCase__=True , ): super().__init__() _lowerCamelCase = layers_per_block _lowerCamelCase = torch.nn.Convad( lowerCamelCase__ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) # down _lowerCamelCase = block_out_channels[0] for i, down_block_type in enumerate(lowerCamelCase__ ): _lowerCamelCase = output_channel _lowerCamelCase = block_out_channels[i] _lowerCamelCase = i == len(lowerCamelCase__ ) - 1 _lowerCamelCase = get_down_block( lowerCamelCase__ , num_layers=self.layers_per_block , in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , add_downsample=not is_final_block , resnet_eps=1e-6 , downsample_padding=0 , resnet_act_fn=lowerCamelCase__ , resnet_groups=lowerCamelCase__ , attention_head_dim=lowerCamelCase__ , temb_channels=lowerCamelCase__ , ) self.down_blocks.append(lowerCamelCase__ ) # mid _lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=lowerCamelCase__ , output_scale_factor=1 , resnet_time_scale_shift='''default''' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCamelCase__ , temb_channels=lowerCamelCase__ , ) # out _lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCamelCase__ , eps=1e-6 ) _lowerCamelCase = nn.SiLU() _lowerCamelCase = 2 * out_channels if double_z else out_channels _lowerCamelCase = nn.Convad(block_out_channels[-1] , lowerCamelCase__ , 3 , padding=1 ) _lowerCamelCase = False def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = x _lowerCamelCase = self.conv_in(lowerCamelCase__ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCamelCase__ ): def custom_forward(*lowerCamelCase__ ): return module(*lowerCamelCase__ ) return custom_forward # down if is_torch_version('''>=''' , '''1.11.0''' ): for down_block in self.down_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) else: for down_block in self.down_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ ) # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCamelCase__ ) else: # down for down_block in self.down_blocks: _lowerCamelCase = down_block(lowerCamelCase__ ) # middle _lowerCamelCase = self.mid_block(lowerCamelCase__ ) # post-process _lowerCamelCase = self.conv_norm_out(lowerCamelCase__ ) _lowerCamelCase = self.conv_act(lowerCamelCase__ ) _lowerCamelCase = self.conv_out(lowerCamelCase__ ) return sample class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__=3 , lowerCamelCase__=3 , lowerCamelCase__=("UpDecoderBlock2D",) , lowerCamelCase__=(6_4,) , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__="silu" , lowerCamelCase__="group" , ): super().__init__() _lowerCamelCase = layers_per_block _lowerCamelCase = nn.Convad( lowerCamelCase__ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _lowerCamelCase = None _lowerCamelCase = nn.ModuleList([] ) _lowerCamelCase = in_channels if norm_type == '''spatial''' else None # mid _lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=lowerCamelCase__ , output_scale_factor=1 , resnet_time_scale_shift='''default''' if norm_type == '''group''' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCamelCase__ , temb_channels=lowerCamelCase__ , ) # up _lowerCamelCase = list(reversed(lowerCamelCase__ ) ) _lowerCamelCase = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCamelCase__ ): _lowerCamelCase = output_channel _lowerCamelCase = reversed_block_out_channels[i] _lowerCamelCase = i == len(lowerCamelCase__ ) - 1 _lowerCamelCase = get_up_block( lowerCamelCase__ , num_layers=self.layers_per_block + 1 , in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , prev_output_channel=lowerCamelCase__ , add_upsample=not is_final_block , resnet_eps=1e-6 , resnet_act_fn=lowerCamelCase__ , resnet_groups=lowerCamelCase__ , attention_head_dim=lowerCamelCase__ , temb_channels=lowerCamelCase__ , resnet_time_scale_shift=lowerCamelCase__ , ) self.up_blocks.append(lowerCamelCase__ ) _lowerCamelCase = output_channel # out if norm_type == "spatial": _lowerCamelCase = SpatialNorm(block_out_channels[0] , lowerCamelCase__ ) else: _lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCamelCase__ , eps=1e-6 ) _lowerCamelCase = nn.SiLU() _lowerCamelCase = nn.Convad(block_out_channels[0] , lowerCamelCase__ , 3 , padding=1 ) _lowerCamelCase = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = z _lowerCamelCase = self.conv_in(lowerCamelCase__ ) _lowerCamelCase = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCamelCase__ ): def custom_forward(*lowerCamelCase__ ): return module(*lowerCamelCase__ ) return custom_forward if is_torch_version('''>=''' , '''1.11.0''' ): # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase__ , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) _lowerCamelCase = sample.to(lowerCamelCase__ ) # up for up_block in self.up_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ , lowerCamelCase__ , use_reentrant=lowerCamelCase__ ) else: # middle _lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = sample.to(lowerCamelCase__ ) # up for up_block in self.up_blocks: _lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCamelCase__ ) , lowerCamelCase__ , lowerCamelCase__ ) else: # middle _lowerCamelCase = self.mid_block(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = sample.to(lowerCamelCase__ ) # up for up_block in self.up_blocks: _lowerCamelCase = up_block(lowerCamelCase__ , lowerCamelCase__ ) # post-process if latent_embeds is None: _lowerCamelCase = self.conv_norm_out(lowerCamelCase__ ) else: _lowerCamelCase = self.conv_norm_out(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.conv_act(lowerCamelCase__ ) _lowerCamelCase = self.conv_out(lowerCamelCase__ ) return sample class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__="random" , lowerCamelCase__=False , lowerCamelCase__=True ): super().__init__() _lowerCamelCase = n_e _lowerCamelCase = vq_embed_dim _lowerCamelCase = beta _lowerCamelCase = legacy _lowerCamelCase = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _lowerCamelCase = remap if self.remap is not None: self.register_buffer('''used''' , torch.tensor(np.load(self.remap ) ) ) _lowerCamelCase = self.used.shape[0] _lowerCamelCase = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _lowerCamelCase = self.re_embed _lowerCamelCase = self.re_embed + 1 print( F"""Remapping {self.n_e} indices to {self.re_embed} indices. """ F"""Using {self.unknown_index} for unknown indices.""" ) else: _lowerCamelCase = n_e _lowerCamelCase = sane_index_shape def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = inds.shape assert len(lowerCamelCase__ ) > 1 _lowerCamelCase = inds.reshape(ishape[0] , -1 ) _lowerCamelCase = self.used.to(lowerCamelCase__ ) _lowerCamelCase = (inds[:, :, None] == used[None, None, ...]).long() _lowerCamelCase = match.argmax(-1 ) _lowerCamelCase = match.sum(2 ) < 1 if self.unknown_index == "random": _lowerCamelCase = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _lowerCamelCase = self.unknown_index return new.reshape(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = inds.shape assert len(lowerCamelCase__ ) > 1 _lowerCamelCase = inds.reshape(ishape[0] , -1 ) _lowerCamelCase = self.used.to(lowerCamelCase__ ) if self.re_embed > self.used.shape[0]: # extra token _lowerCamelCase = 0 # simply set to zero _lowerCamelCase = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCamelCase__ ) return back.reshape(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): # reshape z -> (batch, height, width, channel) and flatten _lowerCamelCase = z.permute(0 , 2 , 3 , 1 ).contiguous() _lowerCamelCase = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _lowerCamelCase = torch.argmin(torch.cdist(lowerCamelCase__ , self.embedding.weight ) , dim=1 ) _lowerCamelCase = self.embedding(lowerCamelCase__ ).view(z.shape ) _lowerCamelCase = None _lowerCamelCase = None # compute loss for embedding if not self.legacy: _lowerCamelCase = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _lowerCamelCase = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _lowerCamelCase = z + (z_q - z).detach() # reshape back to match original input shape _lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _lowerCamelCase = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _lowerCamelCase = self.remap_to_used(lowerCamelCase__ ) _lowerCamelCase = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _lowerCamelCase = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): # shape specifying (batch, height, width, channel) if self.remap is not None: _lowerCamelCase = indices.reshape(shape[0] , -1 ) # add batch axis _lowerCamelCase = self.unmap_to_all(lowerCamelCase__ ) _lowerCamelCase = indices.reshape(-1 ) # flatten again # get quantized latent vectors _lowerCamelCase = self.embedding(lowerCamelCase__ ) if shape is not None: _lowerCamelCase = z_q.view(lowerCamelCase__ ) # reshape back to match original input shape _lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = parameters _lowerCamelCase , _lowerCamelCase = torch.chunk(lowerCamelCase__ , 2 , dim=1 ) _lowerCamelCase = torch.clamp(self.logvar , -3_0.0 , 2_0.0 ) _lowerCamelCase = deterministic _lowerCamelCase = torch.exp(0.5 * self.logvar ) _lowerCamelCase = torch.exp(self.logvar ) if self.deterministic: _lowerCamelCase = _lowerCamelCase = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def snake_case__ ( self , lowerCamelCase__ = None ): # make sure sample is on the same device as the parameters and has same dtype _lowerCamelCase = randn_tensor( self.mean.shape , generator=lowerCamelCase__ , device=self.parameters.device , dtype=self.parameters.dtype ) _lowerCamelCase = self.mean + self.std * sample return x def snake_case__ ( self , lowerCamelCase__=None ): if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=[1, 2, 3] ): if self.deterministic: return torch.Tensor([0.0] ) _lowerCamelCase = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCamelCase__ ) def snake_case__ ( self ): return self.mean
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 1_0: '''a''', 1_1: '''b''', 1_2: '''c''', 1_3: '''d''', 1_4: '''e''', 1_5: '''f''', } def lowerCAmelCase_( lowercase_ : float ) -> str: assert type(lowercase_ ) in (int, float) and decimal == int(lowercase_ ) _lowerCamelCase = int(lowercase_ ) _lowerCamelCase = '''''' _lowerCamelCase = False if decimal < 0: _lowerCamelCase = True decimal *= -1 while decimal > 0: _lowerCamelCase , _lowerCamelCase = divmod(lowercase_ , 16 ) _lowerCamelCase = values[remainder] + hexadecimal _lowerCamelCase = '''0x''' + hexadecimal if negative: _lowerCamelCase = '''-''' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
1
"""simple docstring""" import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration __SCREAMING_SNAKE_CASE : List[str] = 5_0_0_0_0_0 __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.split(__file__) __SCREAMING_SNAKE_CASE : Any = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCAmelCase_( lowercase_ : datasets.Dataset , **lowercase_ : Union[str, Any] ) -> List[str]: _lowerCamelCase = dataset.map(**lowercase_ ) @get_duration def lowerCAmelCase_( lowercase_ : datasets.Dataset , **lowercase_ : Optional[int] ) -> Any: _lowerCamelCase = dataset.filter(**lowercase_ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = {'''num examples''': SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase = datasets.Features({'''text''': datasets.Value('''string''' ), '''numbers''': datasets.Value('''float32''' )} ) _lowerCamelCase = generate_example_dataset( os.path.join(lowercase_ , '''dataset.arrow''' ) , lowercase_ , num_examples=lowercase_ ) _lowerCamelCase = transformers.AutoTokenizer.from_pretrained('''bert-base-cased''' , use_fast=lowercase_ ) def tokenize(lowercase_ : Dict ): return tokenizer(examples['''text'''] ) _lowerCamelCase = map(lowercase_ ) _lowerCamelCase = map(lowercase_ , batched=lowercase_ ) _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''numpy''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''pandas''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''torch''' , columns='''numbers''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) with dataset.formatted_as(type='''tensorflow''' , columns='''numbers''' ): _lowerCamelCase = map(lowercase_ , function=lambda lowercase_ : None , batched=lowercase_ ) _lowerCamelCase = map(lowercase_ , function=lowercase_ , batched=lowercase_ ) _lowerCamelCase = filter(lowercase_ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(lowercase_ , '''wb''' ) as f: f.write(json.dumps(lowercase_ ).encode('''utf-8''' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
623
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) def lowerCAmelCase_( lowercase_ : int ) -> Optional[int]: _lowerCamelCase = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: _lowerCamelCase = [1_44, 1_92, 2_40] _lowerCamelCase = [16, 32, 64, 96, 1_28, 1_60, 6_40] elif "mobilevit_xs" in mobilevit_name: _lowerCamelCase = [96, 1_20, 1_44] _lowerCamelCase = [16, 32, 48, 64, 80, 96, 3_84] elif "mobilevit_xxs" in mobilevit_name: _lowerCamelCase = [64, 80, 96] _lowerCamelCase = [16, 16, 24, 48, 64, 80, 3_20] _lowerCamelCase = 0.0_5 _lowerCamelCase = 2.0 if mobilevit_name.startswith('''deeplabv3_''' ): _lowerCamelCase = 5_12 _lowerCamelCase = 16 _lowerCamelCase = 21 _lowerCamelCase = '''pascal-voc-id2label.json''' else: _lowerCamelCase = 10_00 _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = json.load(open(hf_hub_download(lowercase_ , lowercase_ , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(lowercase_ ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : Union[str, Any]=False ) -> str: for i in range(1 , 6 ): if F"""layer_{i}.""" in name: _lowerCamelCase = name.replace(F"""layer_{i}.""" , F"""encoder.layer.{i - 1}.""" ) if "conv_1." in name: _lowerCamelCase = name.replace('''conv_1.''' , '''conv_stem.''' ) if ".block." in name: _lowerCamelCase = name.replace('''.block.''' , '''.''' ) if "exp_1x1" in name: _lowerCamelCase = name.replace('''exp_1x1''' , '''expand_1x1''' ) if "red_1x1" in name: _lowerCamelCase = name.replace('''red_1x1''' , '''reduce_1x1''' ) if ".local_rep.conv_3x3." in name: _lowerCamelCase = name.replace('''.local_rep.conv_3x3.''' , '''.conv_kxk.''' ) if ".local_rep.conv_1x1." in name: _lowerCamelCase = name.replace('''.local_rep.conv_1x1.''' , '''.conv_1x1.''' ) if ".norm." in name: _lowerCamelCase = name.replace('''.norm.''' , '''.normalization.''' ) if ".conv." in name: _lowerCamelCase = name.replace('''.conv.''' , '''.convolution.''' ) if ".conv_proj." in name: _lowerCamelCase = name.replace('''.conv_proj.''' , '''.conv_projection.''' ) for i in range(0 , 2 ): for j in range(0 , 4 ): if F""".{i}.{j}.""" in name: _lowerCamelCase = name.replace(F""".{i}.{j}.""" , F""".{i}.layer.{j}.""" ) for i in range(2 , 6 ): for j in range(0 , 4 ): if F""".{i}.{j}.""" in name: _lowerCamelCase = name.replace(F""".{i}.{j}.""" , F""".{i}.""" ) if "expand_1x1" in name: _lowerCamelCase = name.replace('''expand_1x1''' , '''downsampling_layer.expand_1x1''' ) if "conv_3x3" in name: _lowerCamelCase = name.replace('''conv_3x3''' , '''downsampling_layer.conv_3x3''' ) if "reduce_1x1" in name: _lowerCamelCase = name.replace('''reduce_1x1''' , '''downsampling_layer.reduce_1x1''' ) for i in range(2 , 5 ): if F""".global_rep.{i}.weight""" in name: _lowerCamelCase = name.replace(F""".global_rep.{i}.weight""" , '''.layernorm.weight''' ) if F""".global_rep.{i}.bias""" in name: _lowerCamelCase = name.replace(F""".global_rep.{i}.bias""" , '''.layernorm.bias''' ) if ".global_rep." in name: _lowerCamelCase = name.replace('''.global_rep.''' , '''.transformer.''' ) if ".pre_norm_mha.0." in name: _lowerCamelCase = name.replace('''.pre_norm_mha.0.''' , '''.layernorm_before.''' ) if ".pre_norm_mha.1.out_proj." in name: _lowerCamelCase = name.replace('''.pre_norm_mha.1.out_proj.''' , '''.attention.output.dense.''' ) if ".pre_norm_ffn.0." in name: _lowerCamelCase = name.replace('''.pre_norm_ffn.0.''' , '''.layernorm_after.''' ) if ".pre_norm_ffn.1." in name: _lowerCamelCase = name.replace('''.pre_norm_ffn.1.''' , '''.intermediate.dense.''' ) if ".pre_norm_ffn.4." in name: _lowerCamelCase = name.replace('''.pre_norm_ffn.4.''' , '''.output.dense.''' ) if ".transformer." in name: _lowerCamelCase = name.replace('''.transformer.''' , '''.transformer.layer.''' ) if ".aspp_layer." in name: _lowerCamelCase = name.replace('''.aspp_layer.''' , '''.''' ) if ".aspp_pool." in name: _lowerCamelCase = name.replace('''.aspp_pool.''' , '''.''' ) if "seg_head." in name: _lowerCamelCase = name.replace('''seg_head.''' , '''segmentation_head.''' ) if "segmentation_head.classifier.classifier." in name: _lowerCamelCase = name.replace('''segmentation_head.classifier.classifier.''' , '''segmentation_head.classifier.''' ) if "classifier.fc." in name: _lowerCamelCase = name.replace('''classifier.fc.''' , '''classifier.''' ) elif (not base_model) and ("segmentation_head." not in name): _lowerCamelCase = '''mobilevit.''' + name return name def lowerCAmelCase_( lowercase_ : str , lowercase_ : List[str] , lowercase_ : List[str]=False ) -> List[Any]: if base_model: _lowerCamelCase = '''''' else: _lowerCamelCase = '''mobilevit.''' for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(lowercase_ ) if key[:8] == "encoder.": _lowerCamelCase = key[8:] if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[0][6:] ) - 1 _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = model.get_submodule(F"""{model_prefix}encoder.layer.{layer_num}""" ) _lowerCamelCase = layer.transformer.layer[transformer_num].attention.attention.all_head_size _lowerCamelCase = ( F"""{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.""" ) if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] else: _lowerCamelCase = val return orig_state_dict def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ) return im @torch.no_grad() def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : Optional[Any]=False ) -> List[Any]: _lowerCamelCase = get_mobilevit_config(lowercase_ ) # load original state_dict _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # load 🤗 model if mobilevit_name.startswith('''deeplabv3_''' ): _lowerCamelCase = MobileViTForSemanticSegmentation(lowercase_ ).eval() else: _lowerCamelCase = MobileViTForImageClassification(lowercase_ ).eval() _lowerCamelCase = convert_state_dict(lowercase_ , lowercase_ ) model.load_state_dict(lowercase_ ) # Check outputs on an image, prepared by MobileViTImageProcessor _lowerCamelCase = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) _lowerCamelCase = image_processor(images=prepare_img() , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) _lowerCamelCase = outputs.logits if mobilevit_name.startswith('''deeplabv3_''' ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": _lowerCamelCase = torch.tensor( [ [[6.2_0_6_5, 6.1_2_9_2, 6.2_0_7_0], [6.1_0_7_9, 6.1_2_5_4, 6.1_7_4_7], [6.0_0_4_2, 6.1_0_7_1, 6.1_0_3_4]], [[-6.9_2_5_3, -6.8_6_5_3, -7.0_3_9_8], [-7.3_2_1_8, -7.3_9_8_3, -7.3_6_7_0], [-7.1_9_6_1, -7.2_4_8_2, -7.1_5_6_9]], [[-4.4_7_2_3, -4.4_3_4_8, -4.3_7_6_9], [-5.3_6_2_9, -5.4_6_3_2, -5.4_5_9_8], [-5.1_5_8_7, -5.3_4_0_2, -5.5_0_5_9]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": _lowerCamelCase = torch.tensor( [ [[5.4_4_4_9, 5.5_7_3_3, 5.6_3_1_4], [5.1_8_1_5, 5.3_9_3_0, 5.5_9_6_3], [5.1_6_5_6, 5.4_3_3_3, 5.4_8_5_3]], [[-9.4_4_2_3, -9.7_7_6_6, -9.6_7_1_4], [-9.1_5_8_1, -9.5_7_2_0, -9.5_5_1_9], [-9.1_0_0_6, -9.6_4_5_8, -9.5_7_0_3]], [[-7.7_7_2_1, -7.3_7_1_6, -7.1_5_8_3], [-8.4_5_9_9, -8.0_6_2_4, -7.7_9_4_4], [-8.4_1_7_2, -7.8_3_6_6, -7.5_0_2_5]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": _lowerCamelCase = torch.tensor( [ [[6.9_8_1_1, 6.9_7_4_3, 7.3_1_2_3], [7.1_7_7_7, 7.1_9_3_1, 7.3_9_3_8], [7.5_6_3_3, 7.8_0_5_0, 7.8_9_0_1]], [[-1_0.5_5_3_6, -1_0.2_3_3_2, -1_0.2_9_2_4], [-1_0.2_3_3_6, -9.8_6_2_4, -9.5_9_6_4], [-1_0.8_8_4_0, -1_0.8_1_5_8, -1_0.6_6_5_9]], [[-3.4_9_3_8, -3.0_6_3_1, -2.8_6_2_0], [-3.4_2_0_5, -2.8_1_3_5, -2.6_8_7_5], [-3.4_1_7_9, -2.7_9_4_5, -2.8_7_5_0]], ] ) else: raise ValueError(F"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3, :3, :3] , lowercase_ , atol=1e-4 ) else: assert logits.shape == (1, 10_00) if mobilevit_name == "mobilevit_s": _lowerCamelCase = torch.tensor([-0.9_8_6_6, 0.2_3_9_2, -1.1_2_4_1] ) elif mobilevit_name == "mobilevit_xs": _lowerCamelCase = torch.tensor([-2.4_7_6_1, -0.9_3_9_9, -1.9_5_8_7] ) elif mobilevit_name == "mobilevit_xxs": _lowerCamelCase = torch.tensor([-1.9_3_6_4, -1.2_3_2_7, -0.4_6_5_3] ) else: raise ValueError(F"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3] , lowercase_ , atol=1e-4 ) Path(lowercase_ ).mkdir(exist_ok=lowercase_ ) print(F"""Saving model {mobilevit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase_ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowercase_ ) if push_to_hub: _lowerCamelCase = { '''mobilevit_s''': '''mobilevit-small''', '''mobilevit_xs''': '''mobilevit-x-small''', '''mobilevit_xxs''': '''mobilevit-xx-small''', '''deeplabv3_mobilevit_s''': '''deeplabv3-mobilevit-small''', '''deeplabv3_mobilevit_xs''': '''deeplabv3-mobilevit-x-small''', '''deeplabv3_mobilevit_xxs''': '''deeplabv3-mobilevit-xx-small''', } print('''Pushing to the hub...''' ) _lowerCamelCase = model_mapping[mobilevit_name] image_processor.push_to_hub(lowercase_ , organization='''apple''' ) model.push_to_hub(lowercase_ , organization='''apple''' ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--mobilevit_name''', default='''mobilevit_s''', type=str, help=( '''Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',''' ''' \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.''' ), ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" from collections import Counter from timeit import timeit def lowerCAmelCase_( lowercase_ : str = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(''' ''' , '''''' ).lower() ).values() ) < 2 def lowerCAmelCase_( lowercase_ : str = "" ) -> bool: if len(lowercase_ ) == 0: return True _lowerCamelCase = input_str.replace(''' ''' , '''''' ).lower() # character_freq_dict: Stores the frequency of every character in the input string _lowerCamelCase = {} for character in lower_case_input_str: _lowerCamelCase = character_freq_dict.get(lowercase_ , 0 ) + 1 _lowerCamelCase = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def lowerCAmelCase_( lowercase_ : str = "" ) -> None: print('''\nFor string = ''' , lowercase_ , ''':''' ) print( '''> can_string_be_rearranged_as_palindrome_counter()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome_counter(lowercase_ ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome_counter(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) print( '''> can_string_be_rearranged_as_palindrome()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome(lowercase_ ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Tuple = input( '''Enter string to determine if it can be rearranged as a palindrome or not: ''' ).strip() benchmark(check_str) __SCREAMING_SNAKE_CASE : int = can_string_be_rearranged_as_palindrome_counter(check_str) print(F"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def lowerCAmelCase_( lowercase_ : Dataset , lowercase_ : Dict[str, str] ) -> Tuple: _lowerCamelCase = args.log_outputs _lowerCamelCase = '''_'''.join(args.dataset.split('''/''' ) + [args.config, args.split] ) # load metric _lowerCamelCase = load_metric('''wer''' ) _lowerCamelCase = load_metric('''cer''' ) # compute metrics _lowerCamelCase = wer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) _lowerCamelCase = cer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) # print & log results _lowerCamelCase = F"""WER: {wer_result}\nCER: {cer_result}""" print(lowercase_ ) with open(F"""{dataset_id}_eval_results.txt""" , '''w''' ) as f: f.write(lowercase_ ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: _lowerCamelCase = F"""log_{dataset_id}_predictions.txt""" _lowerCamelCase = F"""log_{dataset_id}_targets.txt""" with open(lowercase_ , '''w''' ) as p, open(lowercase_ , '''w''' ) as t: # mapping function to write output def write_to_file(lowercase_ : Optional[int] , lowercase_ : List[Any] ): p.write(F"""{i}""" + '''\n''' ) p.write(batch['''prediction'''] + '''\n''' ) t.write(F"""{i}""" + '''\n''' ) t.write(batch['''target'''] + '''\n''' ) result.map(lowercase_ , with_indices=lowercase_ ) def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase = '''[,?.!\-\;\:"“%‘”�—’…–]''' # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training _lowerCamelCase = re.sub(lowercase_ , '''''' , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! _lowerCamelCase = ['''\n\n''', '''\n''', ''' ''', ''' '''] for t in token_sequences_to_ignore: _lowerCamelCase = ''' '''.join(text.split(lowercase_ ) ) return text def lowerCAmelCase_( lowercase_ : Any ) -> List[Any]: # load dataset _lowerCamelCase = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=lowercase_ ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor _lowerCamelCase = AutoFeatureExtractor.from_pretrained(args.model_id ) _lowerCamelCase = feature_extractor.sampling_rate # resample audio _lowerCamelCase = dataset.cast_column('''audio''' , Audio(sampling_rate=lowercase_ ) ) # load eval pipeline if args.device is None: _lowerCamelCase = 0 if torch.cuda.is_available() else -1 _lowerCamelCase = pipeline('''automatic-speech-recognition''' , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(lowercase_ : Optional[Any] ): _lowerCamelCase = asr( batch['''audio''']['''array'''] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) _lowerCamelCase = prediction['''text'''] _lowerCamelCase = normalize_text(batch['''sentence'''] ) return batch # run inference on all examples _lowerCamelCase = dataset.map(lowercase_ , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(lowercase_ , lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--model_id''', type=str, required=True, help='''Model identifier. Should be loadable with 🤗 Transformers''' ) parser.add_argument( '''--dataset''', type=str, required=True, help='''Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets''', ) parser.add_argument( '''--config''', type=str, required=True, help='''Config of the dataset. *E.g.* `\'en\'` for Common Voice''' ) parser.add_argument('''--split''', type=str, required=True, help='''Split of the dataset. *E.g.* `\'test\'`''') parser.add_argument( '''--chunk_length_s''', type=float, default=None, help='''Chunk length in seconds. Defaults to 5 seconds.''' ) parser.add_argument( '''--stride_length_s''', type=float, default=None, help='''Stride of the audio chunks. Defaults to 1 second.''' ) parser.add_argument( '''--log_outputs''', action='''store_true''', help='''If defined, write outputs to log file for analysis.''' ) parser.add_argument( '''--device''', type=int, default=None, help='''The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.''', ) __SCREAMING_SNAKE_CASE : Any = parser.parse_args() main(args)
623
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_inverse_sqrt_schedule, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Any=10 ) -> Optional[Any]: _lowerCamelCase = [] for _ in range(lowercase_ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() return lrs def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any]=10 ) -> List[str]: _lowerCamelCase = [] for step in range(lowercase_ ): lrs.append(scheduler.get_lr()[0] ) scheduler.step() if step == num_steps // 2: with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowercase_ , '''schedule.bin''' ) torch.save(scheduler.state_dict() , lowercase_ ) _lowerCamelCase = torch.load(lowercase_ ) scheduler.load_state_dict(lowercase_ ) return lrs @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for a, b in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertAlmostEqual(lowerCamelCase__ , lowerCamelCase__ , delta=lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=lowerCamelCase__ ) _lowerCamelCase = torch.tensor([0.4, 0.2, -0.5] ) _lowerCamelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _lowerCamelCase = AdamW(params=[w] , lr=2e-1 , weight_decay=0.0 ) for _ in range(1_0_0 ): _lowerCamelCase = criterion(lowerCamelCase__ , lowerCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) def snake_case__ ( self ): _lowerCamelCase = torch.tensor([0.1, -0.2, -0.1] , requires_grad=lowerCamelCase__ ) _lowerCamelCase = torch.tensor([0.4, 0.2, -0.5] ) _lowerCamelCase = nn.MSELoss() # No warmup, constant schedule, no gradient clipping _lowerCamelCase = Adafactor( params=[w] , lr=1e-2 , eps=(1e-30, 1e-3) , clip_threshold=1.0 , decay_rate=-0.8 , betaa=lowerCamelCase__ , weight_decay=0.0 , relative_step=lowerCamelCase__ , scale_parameter=lowerCamelCase__ , warmup_init=lowerCamelCase__ , ) for _ in range(1_0_0_0 ): _lowerCamelCase = criterion(lowerCamelCase__ , lowerCamelCase__ ) loss.backward() optimizer.step() w.grad.detach_() # No zero_grad() function on simple tensors. we do it ourselves. w.grad.zero_() self.assertListAlmostEqual(w.tolist() , [0.4, 0.2, -0.5] , tol=1e-2 ) @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = nn.Linear(50, 50 ) if is_torch_available() else None lowercase__ : int = AdamW(m.parameters(), lr=10.0 ) if is_torch_available() else None lowercase__ : List[str] = 10 def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None ): self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for a, b in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertAlmostEqual(lowerCamelCase__ , lowerCamelCase__ , delta=lowerCamelCase__ , msg=lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = {'''num_warmup_steps''': 2, '''num_training_steps''': 1_0} # schedulers doct format # function: (sched_args_dict, expected_learning_rates) _lowerCamelCase = { get_constant_schedule: ({}, [1_0.0] * self.num_steps), get_constant_schedule_with_warmup: ( {'''num_warmup_steps''': 4}, [0.0, 2.5, 5.0, 7.5, 1_0.0, 1_0.0, 1_0.0, 1_0.0, 1_0.0, 1_0.0], ), get_linear_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 1_0.0, 8.7_5, 7.5, 6.2_5, 5.0, 3.7_5, 2.5, 1.2_5], ), get_cosine_schedule_with_warmup: ( {**common_kwargs}, [0.0, 5.0, 1_0.0, 9.6_1, 8.5_3, 6.9_1, 5.0, 3.0_8, 1.4_6, 0.3_8], ), get_cosine_with_hard_restarts_schedule_with_warmup: ( {**common_kwargs, '''num_cycles''': 2}, [0.0, 5.0, 1_0.0, 8.5_3, 5.0, 1.4_6, 1_0.0, 8.5_3, 5.0, 1.4_6], ), get_polynomial_decay_schedule_with_warmup: ( {**common_kwargs, '''power''': 2.0, '''lr_end''': 1e-7}, [0.0, 5.0, 1_0.0, 7.6_5_6, 5.6_2_5, 3.9_0_6, 2.5, 1.4_0_6, 0.6_2_5, 0.1_5_6], ), get_inverse_sqrt_schedule: ( {'''num_warmup_steps''': 2}, [0.0, 5.0, 1_0.0, 8.1_6_5, 7.0_7_1, 6.3_2_5, 5.7_7_4, 5.3_4_5, 5.0, 4.7_1_4], ), } for scheduler_func, data in scheds.items(): _lowerCamelCase , _lowerCamelCase = data _lowerCamelCase = scheduler_func(self.optimizer , **lowerCamelCase__ ) self.assertEqual(len([scheduler.get_lr()[0]] ) , 1 ) _lowerCamelCase = unwrap_schedule(lowerCamelCase__ , self.num_steps ) self.assertListAlmostEqual( lowerCamelCase__ , lowerCamelCase__ , tol=1e-2 , msg=F"""failed for {scheduler_func} in normal scheduler""" , ) _lowerCamelCase = scheduler_func(self.optimizer , **lowerCamelCase__ ) if scheduler_func.__name__ != "get_constant_schedule": LambdaScheduleWrapper.wrap_scheduler(lowerCamelCase__ ) # wrap to test picklability of the schedule _lowerCamelCase = unwrap_and_save_reload_schedule(lowerCamelCase__ , self.num_steps ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ , msg=F"""failed for {scheduler_func} in save and reload""" ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = fn def __call__( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.fn(*lowerCamelCase__ , **lowerCamelCase__ ) @classmethod def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = list(map(self , scheduler.lr_lambdas ) )
623
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
1
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_( lowercase_ : Any="ro" , lowercase_ : Optional[int]="en" , lowercase_ : int="wmt16" , lowercase_ : int=None ) -> None: try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError('''run pip install datasets''' ) _lowerCamelCase = F"""{src_lang}-{tgt_lang}""" print(F"""Converting {dataset}-{pair}""" ) _lowerCamelCase = datasets.load_dataset(lowercase_ , lowercase_ ) if save_dir is None: _lowerCamelCase = F"""{dataset}-{pair}""" _lowerCamelCase = Path(lowercase_ ) save_dir.mkdir(exist_ok=lowercase_ ) for split in ds.keys(): print(F"""Splitting {split} with {ds[split].num_rows} records""" ) # to save to val.source, val.target like summary datasets _lowerCamelCase = '''val''' if split == '''validation''' else split _lowerCamelCase = save_dir.joinpath(F"""{fn}.source""" ) _lowerCamelCase = save_dir.joinpath(F"""{fn}.target""" ) _lowerCamelCase = src_path.open('''w+''' ) _lowerCamelCase = tgt_path.open('''w+''' ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): _lowerCamelCase = x['''translation'''] src_fp.write(ex[src_lang] + '''\n''' ) tgt_fp.write(ex[tgt_lang] + '''\n''' ) print(F"""Saved {dataset} dataset to {save_dir}""" ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
623
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 10_00 ) -> int: _lowerCamelCase , _lowerCamelCase = 1, 1 _lowerCamelCase = 2 while True: _lowerCamelCase = 0 _lowerCamelCase = fa + fa _lowerCamelCase , _lowerCamelCase = fa, f index += 1 for _ in str(lowercase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
623
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
623
1
"""simple docstring""" import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) def lowerCAmelCase_( lowercase_ : Tuple ) -> Any: _lowerCamelCase = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: _lowerCamelCase = 1_28 elif "12-12" in model_name: _lowerCamelCase = 12 _lowerCamelCase = 12 elif "14-14" in model_name: _lowerCamelCase = 14 _lowerCamelCase = 14 elif "16-16" in model_name: _lowerCamelCase = 16 _lowerCamelCase = 16 else: raise ValueError('''Model not supported''' ) _lowerCamelCase = '''huggingface/label-files''' if "speech-commands" in model_name: _lowerCamelCase = 35 _lowerCamelCase = '''speech-commands-v2-id2label.json''' else: _lowerCamelCase = 5_27 _lowerCamelCase = '''audioset-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(lowercase_ , lowercase_ , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(lowercase_ ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Tuple: if "module.v" in name: _lowerCamelCase = name.replace('''module.v''' , '''audio_spectrogram_transformer''' ) if "cls_token" in name: _lowerCamelCase = name.replace('''cls_token''' , '''embeddings.cls_token''' ) if "dist_token" in name: _lowerCamelCase = name.replace('''dist_token''' , '''embeddings.distillation_token''' ) if "pos_embed" in name: _lowerCamelCase = name.replace('''pos_embed''' , '''embeddings.position_embeddings''' ) if "patch_embed.proj" in name: _lowerCamelCase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) # transformer blocks if "blocks" in name: _lowerCamelCase = name.replace('''blocks''' , '''encoder.layer''' ) if "attn.proj" in name: _lowerCamelCase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: _lowerCamelCase = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: _lowerCamelCase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: _lowerCamelCase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: _lowerCamelCase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: _lowerCamelCase = name.replace('''mlp.fc2''' , '''output.dense''' ) # final layernorm if "audio_spectrogram_transformer.norm" in name: _lowerCamelCase = name.replace('''audio_spectrogram_transformer.norm''' , '''audio_spectrogram_transformer.layernorm''' ) # classifier head if "module.mlp_head.0" in name: _lowerCamelCase = name.replace('''module.mlp_head.0''' , '''classifier.layernorm''' ) if "module.mlp_head.1" in name: _lowerCamelCase = name.replace('''module.mlp_head.1''' , '''classifier.dense''' ) return name def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Any ) -> str: for key in orig_state_dict.copy().keys(): _lowerCamelCase = orig_state_dict.pop(lowercase_ ) if "qkv" in key: _lowerCamelCase = key.split('''.''' ) _lowerCamelCase = int(key_split[3] ) _lowerCamelCase = config.hidden_size if "weight" in key: _lowerCamelCase = val[:dim, :] _lowerCamelCase = val[dim : dim * 2, :] _lowerCamelCase = val[-dim:, :] else: _lowerCamelCase = val[:dim] _lowerCamelCase = val[dim : dim * 2] _lowerCamelCase = val[-dim:] else: _lowerCamelCase = val return orig_state_dict def lowerCAmelCase_( lowercase_ : Any ) -> Tuple: _lowerCamelCase = [ '''module.v.head.weight''', '''module.v.head.bias''', '''module.v.head_dist.weight''', '''module.v.head_dist.bias''', ] for k in ignore_keys: state_dict.pop(lowercase_ , lowercase_ ) @torch.no_grad() def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : Any=False ) -> Optional[Any]: _lowerCamelCase = get_audio_spectrogram_transformer_config(lowercase_ ) _lowerCamelCase = { '''ast-finetuned-audioset-10-10-0.4593''': ( '''https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1''' ), '''ast-finetuned-audioset-10-10-0.450''': ( '''https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1''' ), '''ast-finetuned-audioset-10-10-0.448''': ( '''https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1''' ), '''ast-finetuned-audioset-10-10-0.448-v2''': ( '''https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1''' ), '''ast-finetuned-audioset-12-12-0.447''': ( '''https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1''' ), '''ast-finetuned-audioset-14-14-0.443''': ( '''https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1''' ), '''ast-finetuned-audioset-16-16-0.442''': ( '''https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1''' ), '''ast-finetuned-speech-commands-v2''': ( '''https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1''' ), } # load original state_dict _lowerCamelCase = model_name_to_url[model_name] _lowerCamelCase = torch.hub.load_state_dict_from_url(lowercase_ , map_location='''cpu''' ) # remove some keys remove_keys(lowercase_ ) # rename some keys _lowerCamelCase = convert_state_dict(lowercase_ , lowercase_ ) # load 🤗 model _lowerCamelCase = ASTForAudioClassification(lowercase_ ) model.eval() model.load_state_dict(lowercase_ ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 _lowerCamelCase = -4.2_6_7_7_3_9_3 if '''speech-commands''' not in model_name else -6.8_4_5_9_7_8 _lowerCamelCase = 4.5_6_8_9_9_7_4 if '''speech-commands''' not in model_name else 5.5_6_5_4_5_2_6 _lowerCamelCase = 10_24 if '''speech-commands''' not in model_name else 1_28 _lowerCamelCase = ASTFeatureExtractor(mean=lowercase_ , std=lowercase_ , max_length=lowercase_ ) if "speech-commands" in model_name: _lowerCamelCase = load_dataset('''speech_commands''' , '''v0.02''' , split='''validation''' ) _lowerCamelCase = dataset[0]['''audio''']['''array'''] else: _lowerCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' , ) _lowerCamelCase , _lowerCamelCase = torchaudio.load(lowercase_ ) _lowerCamelCase = waveform.squeeze().numpy() _lowerCamelCase = feature_extractor(lowercase_ , sampling_rate=1_60_00 , return_tensors='''pt''' ) # forward pass _lowerCamelCase = model(**lowercase_ ) _lowerCamelCase = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": _lowerCamelCase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": _lowerCamelCase = torch.tensor([-1.1_9_8_6, -7.0_9_0_3, -8.2_7_1_8] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": _lowerCamelCase = torch.tensor([-2.6_1_2_8, -8.0_0_8_0, -9.4_3_4_4] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": _lowerCamelCase = torch.tensor([-1.5_0_8_0, -7.4_5_3_4, -8.8_9_1_7] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": _lowerCamelCase = torch.tensor([-0.5_0_5_0, -6.5_8_3_3, -8.0_8_4_3] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": _lowerCamelCase = torch.tensor([-0.3_8_2_6, -7.0_3_3_6, -8.2_4_1_3] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": _lowerCamelCase = torch.tensor([-1.2_1_1_3, -6.9_1_0_1, -8.3_4_7_0] ) elif model_name == "ast-finetuned-speech-commands-v2": _lowerCamelCase = torch.tensor([6.1_5_8_9, -8.0_5_6_6, -8.7_9_8_4] ) else: raise ValueError('''Unknown model name''' ) if not torch.allclose(logits[0, :3] , lowercase_ , atol=1e-4 ): raise ValueError('''Logits don\'t match''' ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(lowercase_ ).mkdir(exist_ok=lowercase_ ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowercase_ ) print(F"""Saving feature extractor to {pytorch_dump_folder_path}""" ) feature_extractor.save_pretrained(lowercase_ ) if push_to_hub: print('''Pushing model and feature extractor to the hub...''' ) model.push_to_hub(F"""MIT/{model_name}""" ) feature_extractor.push_to_hub(F"""MIT/{model_name}""" ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''ast-finetuned-audioset-10-10-0.4593''', type=str, help='''Name of the Audio Spectrogram Transformer model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
623
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = 8.31_4462 # Unit - J mol-1 K-1 def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: if moles < 0 or kelvin < 0 or volume < 0: raise ValueError('''Invalid inputs. Enter positive value.''' ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> float: if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError('''Invalid inputs. Enter positive value.''' ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
623
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
1
"""simple docstring""" from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # General docstring __SCREAMING_SNAKE_CASE : List[str] = '''RegNetConfig''' # Base docstring __SCREAMING_SNAKE_CASE : Optional[int] = '''facebook/regnet-y-040''' __SCREAMING_SNAKE_CASE : int = [1, 1_0_8_8, 7, 7] # Image classification docstring __SCREAMING_SNAKE_CASE : List[str] = '''facebook/regnet-y-040''' __SCREAMING_SNAKE_CASE : int = '''tabby, tabby cat''' __SCREAMING_SNAKE_CASE : str = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 3 , lowerCamelCase__ = 1 , lowerCamelCase__ = 1 , lowerCamelCase__ = "relu" , ): super().__init__() _lowerCamelCase = nn.Convad( lowerCamelCase__ , lowerCamelCase__ , kernel_size=lowerCamelCase__ , stride=lowerCamelCase__ , padding=kernel_size // 2 , groups=lowerCamelCase__ , bias=lowerCamelCase__ , ) _lowerCamelCase = nn.BatchNormad(lowerCamelCase__ ) _lowerCamelCase = ACTaFN[activation] if activation is not None else nn.Identity() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.convolution(lowerCamelCase__ ) _lowerCamelCase = self.normalization(lowerCamelCase__ ) _lowerCamelCase = self.activation(lowerCamelCase__ ) return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__() _lowerCamelCase = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) _lowerCamelCase = config.num_channels def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) _lowerCamelCase = self.embedder(lowerCamelCase__ ) return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 2 ): super().__init__() _lowerCamelCase = nn.Convad(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 , stride=lowerCamelCase__ , bias=lowerCamelCase__ ) _lowerCamelCase = nn.BatchNormad(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.convolution(lowerCamelCase__ ) _lowerCamelCase = self.normalization(lowerCamelCase__ ) return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): super().__init__() _lowerCamelCase = nn.AdaptiveAvgPoolad((1, 1) ) _lowerCamelCase = nn.Sequential( nn.Convad(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 ) , nn.ReLU() , nn.Convad(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 ) , nn.Sigmoid() , ) def snake_case__ ( self , lowerCamelCase__ ): # b c h w -> b c 1 1 _lowerCamelCase = self.pooler(lowerCamelCase__ ) _lowerCamelCase = self.attention(lowerCamelCase__ ) _lowerCamelCase = hidden_state * attention return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 1 ): super().__init__() _lowerCamelCase = in_channels != out_channels or stride != 1 _lowerCamelCase = max(1 , out_channels // config.groups_width ) _lowerCamelCase = ( RegNetShortCut(lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ ) if should_apply_shortcut else nn.Identity() ) _lowerCamelCase = nn.Sequential( RegNetConvLayer(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act ) , RegNetConvLayer(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ ) , ) _lowerCamelCase = ACTaFN[config.hidden_act] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = hidden_state _lowerCamelCase = self.layer(lowerCamelCase__ ) _lowerCamelCase = self.shortcut(lowerCamelCase__ ) hidden_state += residual _lowerCamelCase = self.activation(lowerCamelCase__ ) return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 1 ): super().__init__() _lowerCamelCase = in_channels != out_channels or stride != 1 _lowerCamelCase = max(1 , out_channels // config.groups_width ) _lowerCamelCase = ( RegNetShortCut(lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ ) if should_apply_shortcut else nn.Identity() ) _lowerCamelCase = nn.Sequential( RegNetConvLayer(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , groups=lowerCamelCase__ , activation=config.hidden_act ) , RegNetSELayer(lowerCamelCase__ , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(lowerCamelCase__ , lowerCamelCase__ , kernel_size=1 , activation=lowerCamelCase__ ) , ) _lowerCamelCase = ACTaFN[config.hidden_act] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = hidden_state _lowerCamelCase = self.layer(lowerCamelCase__ ) _lowerCamelCase = self.shortcut(lowerCamelCase__ ) hidden_state += residual _lowerCamelCase = self.activation(lowerCamelCase__ ) return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 2 , lowerCamelCase__ = 2 , ): super().__init__() _lowerCamelCase = RegNetXLayer if config.layer_type == '''x''' else RegNetYLayer _lowerCamelCase = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , stride=lowerCamelCase__ , ) , *[layer(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for _ in range(depth - 1 )] , ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.layers(lowerCamelCase__ ) return hidden_state class lowerCamelCase_( nn.Module ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__() _lowerCamelCase = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( lowerCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) _lowerCamelCase = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(lowerCamelCase__ , config.depths[1:] ): self.stages.append(RegNetStage(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , depth=lowerCamelCase__ ) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False , lowerCamelCase__ = True ): _lowerCamelCase = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _lowerCamelCase = hidden_states + (hidden_state,) _lowerCamelCase = stage_module(lowerCamelCase__ ) if output_hidden_states: _lowerCamelCase = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase__ , hidden_states=lowerCamelCase__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Tuple = RegNetConfig lowercase__ : str = 'regnet' lowercase__ : List[str] = 'pixel_values' lowercase__ : str = True def snake_case__ ( self , lowerCamelCase__ ): if isinstance(lowerCamelCase__ , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode='''fan_out''' , nonlinearity='''relu''' ) elif isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False ): if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = value __SCREAMING_SNAKE_CASE : List[str] = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __SCREAMING_SNAKE_CASE : Any = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( 'The bare RegNet model outputting raw features without any specific head on top.', A__, ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__(lowerCamelCase__ ) _lowerCamelCase = config _lowerCamelCase = RegNetEmbeddings(lowerCamelCase__ ) _lowerCamelCase = RegNetEncoder(lowerCamelCase__ ) _lowerCamelCase = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None ): _lowerCamelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _lowerCamelCase = return_dict if return_dict is not None else self.config.use_return_dict _lowerCamelCase = self.embedder(lowerCamelCase__ ) _lowerCamelCase = self.encoder( lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ ) _lowerCamelCase = encoder_outputs[0] _lowerCamelCase = self.pooler(lowerCamelCase__ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase__ , pooler_output=lowerCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( '\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ', A__, ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__(lowerCamelCase__ ) _lowerCamelCase = config.num_labels _lowerCamelCase = RegNetModel(lowerCamelCase__ ) # classification head _lowerCamelCase = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowerCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def snake_case__ ( self , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , ): _lowerCamelCase = return_dict if return_dict is not None else self.config.use_return_dict _lowerCamelCase = self.regnet(lowerCamelCase__ , output_hidden_states=lowerCamelCase__ , return_dict=lowerCamelCase__ ) _lowerCamelCase = outputs.pooler_output if return_dict else outputs[1] _lowerCamelCase = self.classifier(lowerCamelCase__ ) _lowerCamelCase = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _lowerCamelCase = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _lowerCamelCase = '''single_label_classification''' else: _lowerCamelCase = '''multi_label_classification''' if self.config.problem_type == "regression": _lowerCamelCase = MSELoss() if self.num_labels == 1: _lowerCamelCase = loss_fct(logits.squeeze() , labels.squeeze() ) else: _lowerCamelCase = loss_fct(lowerCamelCase__ , lowerCamelCase__ ) elif self.config.problem_type == "single_label_classification": _lowerCamelCase = CrossEntropyLoss() _lowerCamelCase = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _lowerCamelCase = BCEWithLogitsLoss() _lowerCamelCase = loss_fct(lowerCamelCase__ , lowerCamelCase__ ) if not return_dict: _lowerCamelCase = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=lowerCamelCase__ , logits=lowerCamelCase__ , hidden_states=outputs.hidden_states )
623
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''caidas/swin2sr-classicalsr-x2-64''': ( '''https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json''' ), } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'swin2sr' lowercase__ : Tuple = { 'hidden_size': 'embed_dim', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self , lowerCamelCase__=6_4 , lowerCamelCase__=1 , lowerCamelCase__=3 , lowerCamelCase__=1_8_0 , lowerCamelCase__=[6, 6, 6, 6, 6, 6] , lowerCamelCase__=[6, 6, 6, 6, 6, 6] , lowerCamelCase__=8 , lowerCamelCase__=2.0 , lowerCamelCase__=True , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.1 , lowerCamelCase__="gelu" , lowerCamelCase__=False , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-5 , lowerCamelCase__=2 , lowerCamelCase__=1.0 , lowerCamelCase__="1conv" , lowerCamelCase__="pixelshuffle" , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = embed_dim _lowerCamelCase = depths _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = num_heads _lowerCamelCase = window_size _lowerCamelCase = mlp_ratio _lowerCamelCase = qkv_bias _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = drop_path_rate _lowerCamelCase = hidden_act _lowerCamelCase = use_absolute_embeddings _lowerCamelCase = layer_norm_eps _lowerCamelCase = initializer_range _lowerCamelCase = upscale _lowerCamelCase = img_range _lowerCamelCase = resi_connection _lowerCamelCase = upsampler
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
1
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py __SCREAMING_SNAKE_CASE : List[Any] = '''\ @INPROCEEDINGS{Papineni02bleu:a, author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu}, title = {BLEU: a Method for Automatic Evaluation of Machine Translation}, booktitle = {}, year = {2002}, pages = {311--318} } @inproceedings{lin-och-2004-orange, title = "{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation", author = "Lin, Chin-Yew and Och, Franz Josef", booktitle = "{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics", month = "aug 23{--}aug 27", year = "2004", address = "Geneva, Switzerland", publisher = "COLING", url = "https://www.aclweb.org/anthology/C04-1072", pages = "501--507", } ''' __SCREAMING_SNAKE_CASE : Optional[Any] = '''\ BLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another. Quality is considered to be the correspondence between a machine\'s output and that of a human: "the closer a machine translation is to a professional human translation, the better it is" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and remains one of the most popular automated and inexpensive metrics. Scores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations. Those scores are then averaged over the whole corpus to reach an estimate of the translation\'s overall quality. Intelligibility or grammatical correctness are not taken into account[citation needed]. BLEU\'s output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1 representing more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the reference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional reference translations will increase the BLEU score. ''' __SCREAMING_SNAKE_CASE : str = ''' Computes BLEU score of translated segments against one or more references. Args: predictions: list of translations to score. Each translation should be tokenized into a list of tokens. references: list of lists of references for each translation. Each reference should be tokenized into a list of tokens. max_order: Maximum n-gram order to use when computing BLEU score. smooth: Whether or not to apply Lin et al. 2004 smoothing. Returns: \'bleu\': bleu score, \'precisions\': geometric mean of n-gram precisions, \'brevity_penalty\': brevity penalty, \'length_ratio\': ratio of lengths, \'translation_length\': translation_length, \'reference_length\': reference_length Examples: >>> predictions = [ ... ["hello", "there", "general", "kenobi"], # tokenized prediction of the first sample ... ["foo", "bar", "foobar"] # tokenized prediction of the second sample ... ] >>> references = [ ... [["hello", "there", "general", "kenobi"], ["hello", "there", "!"]], # tokenized references for the first sample (2 references) ... [["foo", "bar", "foobar"]] # tokenized references for the second sample (1 reference) ... ] >>> bleu = datasets.load_metric("bleu") >>> results = bleu.compute(predictions=predictions, references=references) >>> print(results["bleu"]) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ), '''references''': datasets.Sequence( datasets.Sequence(datasets.Value('''string''' , id='''token''' ) , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=['''https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=4 , lowerCamelCase__=False ): _lowerCamelCase = compute_bleu( reference_corpus=lowerCamelCase__ , translation_corpus=lowerCamelCase__ , max_order=lowerCamelCase__ , smooth=lowerCamelCase__ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
623
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : int = { '''configuration_roformer''': ['''ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoFormerConfig''', '''RoFormerOnnxConfig'''], '''tokenization_roformer''': ['''RoFormerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''RoFormerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoFormerForCausalLM''', '''RoFormerForMaskedLM''', '''RoFormerForMultipleChoice''', '''RoFormerForQuestionAnswering''', '''RoFormerForSequenceClassification''', '''RoFormerForTokenClassification''', '''RoFormerLayer''', '''RoFormerModel''', '''RoFormerPreTrainedModel''', '''load_tf_weights_in_roformer''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRoFormerForCausalLM''', '''TFRoFormerForMaskedLM''', '''TFRoFormerForMultipleChoice''', '''TFRoFormerForQuestionAnswering''', '''TFRoFormerForSequenceClassification''', '''TFRoFormerForTokenClassification''', '''TFRoFormerLayer''', '''TFRoFormerModel''', '''TFRoFormerPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ '''FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FlaxRoFormerForMaskedLM''', '''FlaxRoFormerForMultipleChoice''', '''FlaxRoFormerForQuestionAnswering''', '''FlaxRoFormerForSequenceClassification''', '''FlaxRoFormerForTokenClassification''', '''FlaxRoFormerModel''', '''FlaxRoFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
"""simple docstring""" from typing import Any class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) _lowerCamelCase = temp.next print() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = Node(lowerCamelCase__ ) _lowerCamelCase = self.head _lowerCamelCase = new_node def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if node_data_a == node_data_a: return else: _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next if node_a is None or node_a is None: return _lowerCamelCase , _lowerCamelCase = node_a.data, node_a.data if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
623
1
"""simple docstring""" from __future__ import annotations from collections import deque from collections.abc import Iterator from dataclasses import dataclass @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : int lowercase__ : int class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = [[] for _ in range(lowerCamelCase__ )] _lowerCamelCase = size def __getitem__( self , lowerCamelCase__ ): return iter(self._graph[vertex] ) @property def snake_case__ ( self ): return self._size def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if weight not in (0, 1): raise ValueError('''Edge weight must be either 0 or 1.''' ) if to_vertex < 0 or to_vertex >= self.size: raise ValueError('''Vertex indexes must be in [0; size).''' ) self._graph[from_vertex].append(Edge(lowerCamelCase__ , lowerCamelCase__ ) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = deque([start_vertex] ) _lowerCamelCase = [None] * self.size _lowerCamelCase = 0 while queue: _lowerCamelCase = queue.popleft() _lowerCamelCase = distances[current_vertex] if current_distance is None: continue for edge in self[current_vertex]: _lowerCamelCase = current_distance + edge.weight _lowerCamelCase = distances[edge.destination_vertex] if ( isinstance(lowerCamelCase__ , lowerCamelCase__ ) and new_distance >= dest_vertex_distance ): continue _lowerCamelCase = new_distance if edge.weight == 0: queue.appendleft(edge.destination_vertex ) else: queue.append(edge.destination_vertex ) if distances[finish_vertex] is None: raise ValueError('''No path from start_vertex to finish_vertex.''' ) return distances[finish_vertex] if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __SCREAMING_SNAKE_CASE : Optional[Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[Any]: from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> List[str]: from diffusers.utils.testing_utils import pytest_terminal_summary_main _lowerCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
623
1
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[Any] = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Optional[int] = 'van' def __init__( self , lowerCamelCase__=2_2_4 , lowerCamelCase__=3 , lowerCamelCase__=[7, 3, 3, 3] , lowerCamelCase__=[4, 2, 2, 2] , lowerCamelCase__=[6_4, 1_2_8, 3_2_0, 5_1_2] , lowerCamelCase__=[3, 3, 1_2, 3] , lowerCamelCase__=[8, 8, 4, 4] , lowerCamelCase__="gelu" , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-6 , lowerCamelCase__=1e-2 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = image_size _lowerCamelCase = num_channels _lowerCamelCase = patch_sizes _lowerCamelCase = strides _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = mlp_ratios _lowerCamelCase = hidden_act _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = layer_scale_init_value _lowerCamelCase = drop_path_rate _lowerCamelCase = dropout_rate
623
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
1
"""simple docstring""" import socket def lowerCAmelCase_( ) -> Dict: _lowerCamelCase = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase = socket.gethostname() _lowerCamelCase = 1_23_12 sock.connect((host, port) ) sock.send(B'''Hello server!''' ) with open('''Received_file''' , '''wb''' ) as out_file: print('''File opened''' ) print('''Receiving data...''' ) while True: _lowerCamelCase = sock.recv(10_24 ) if not data: break out_file.write(lowercase_ ) print('''Successfully received the file''' ) sock.close() print('''Connection closed''' ) if __name__ == "__main__": main()
623
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
1
"""simple docstring""" import math import sys def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase = '''''' try: with open(lowercase_ , '''rb''' ) as binary_file: _lowerCamelCase = binary_file.read() for dat in data: _lowerCamelCase = F"""{dat:08b}""" result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase = {'''0''': '''0''', '''1''': '''1'''} _lowerCamelCase , _lowerCamelCase = '''''', '''''' _lowerCamelCase = len(lowercase_ ) for i in range(len(lowercase_ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue _lowerCamelCase = lexicon[curr_string] result += last_match_id _lowerCamelCase = last_match_id + '''0''' if math.loga(lowercase_ ).is_integer(): _lowerCamelCase = {} for curr_key in list(lowercase_ ): _lowerCamelCase = lexicon.pop(lowercase_ ) _lowerCamelCase = new_lex _lowerCamelCase = last_match_id + '''1''' index += 1 _lowerCamelCase = '''''' return result def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> None: _lowerCamelCase = 8 try: with open(lowercase_ , '''wb''' ) as opened_file: _lowerCamelCase = [ to_write[i : i + byte_length] for i in range(0 , len(lowercase_ ) , lowercase_ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(lowercase_ , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase = 0 for letter in data_bits: if letter == "1": break counter += 1 _lowerCamelCase = data_bits[counter:] _lowerCamelCase = data_bits[counter + 1 :] return data_bits def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> None: _lowerCamelCase = read_file_binary(lowercase_ ) _lowerCamelCase = remove_prefix(lowercase_ ) _lowerCamelCase = decompress_data(lowercase_ ) write_file_binary(lowercase_ , lowercase_ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
623
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 50 ) -> int: _lowerCamelCase = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
623
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
1
"""simple docstring""" import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } __SCREAMING_SNAKE_CASE : int = { '''b0''': { '''hidden_dim''': 1_2_8_0, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 2_2_4, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1_2_8_0, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 2_4_0, '''dropout_rate''': 0.2, '''dw_padding''': [1_6], }, '''b2''': { '''hidden_dim''': 1_4_0_8, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 2_6_0, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 1_6], }, '''b3''': { '''hidden_dim''': 1_5_3_6, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 3_0_0, '''dropout_rate''': 0.3, '''dw_padding''': [5, 1_8], }, '''b4''': { '''hidden_dim''': 1_7_9_2, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 3_8_0, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2_0_4_8, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 4_5_6, '''dropout_rate''': 0.4, '''dw_padding''': [1_3, 2_7], }, '''b6''': { '''hidden_dim''': 2_3_0_4, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 5_2_8, '''dropout_rate''': 0.5, '''dw_padding''': [3_1], }, '''b7''': { '''hidden_dim''': 2_5_6_0, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 6_0_0, '''dropout_rate''': 0.5, '''dw_padding''': [1_8], }, } def lowerCAmelCase_( lowercase_ : str ) -> Optional[Any]: _lowerCamelCase = EfficientNetConfig() _lowerCamelCase = CONFIG_MAP[model_name]['''hidden_dim'''] _lowerCamelCase = CONFIG_MAP[model_name]['''width_coef'''] _lowerCamelCase = CONFIG_MAP[model_name]['''depth_coef'''] _lowerCamelCase = CONFIG_MAP[model_name]['''image_size'''] _lowerCamelCase = CONFIG_MAP[model_name]['''dropout_rate'''] _lowerCamelCase = CONFIG_MAP[model_name]['''dw_padding'''] _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''imagenet-1k-id2label.json''' _lowerCamelCase = 10_00 _lowerCamelCase = json.load(open(hf_hub_download(lowercase_ , lowercase_ , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(lowercase_ ): v for k, v in idalabel.items()} _lowerCamelCase = idalabel _lowerCamelCase = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase_( ) -> str: _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ) return im def lowerCAmelCase_( lowercase_ : str ) -> Any: _lowerCamelCase = CONFIG_MAP[model_name]['''image_size'''] _lowerCamelCase = EfficientNetImageProcessor( size={'''height''': size, '''width''': size} , image_mean=[0.4_8_5, 0.4_5_6, 0.4_0_6] , image_std=[0.4_7_8_5_3_9_4_4, 0.4_7_3_2_8_6_4, 0.4_7_4_3_4_1_6_3] , do_center_crop=lowercase_ , ) return preprocessor def lowerCAmelCase_( lowercase_ : Any ) -> Dict: _lowerCamelCase = [v.split('''_''' )[0].split('''block''' )[1] for v in original_param_names if v.startswith('''block''' )] _lowerCamelCase = sorted(set(lowercase_ ) ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = {b: str(lowercase_ ) for b, i in zip(lowercase_ , range(lowercase_ ) )} _lowerCamelCase = [] rename_keys.append(('''stem_conv/kernel:0''', '''embeddings.convolution.weight''') ) rename_keys.append(('''stem_bn/gamma:0''', '''embeddings.batchnorm.weight''') ) rename_keys.append(('''stem_bn/beta:0''', '''embeddings.batchnorm.bias''') ) rename_keys.append(('''stem_bn/moving_mean:0''', '''embeddings.batchnorm.running_mean''') ) rename_keys.append(('''stem_bn/moving_variance:0''', '''embeddings.batchnorm.running_var''') ) for b in block_names: _lowerCamelCase = block_name_mapping[b] rename_keys.append((F"""block{b}_expand_conv/kernel:0""", F"""encoder.blocks.{hf_b}.expansion.expand_conv.weight""") ) rename_keys.append((F"""block{b}_expand_bn/gamma:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.weight""") ) rename_keys.append((F"""block{b}_expand_bn/beta:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.bias""") ) rename_keys.append( (F"""block{b}_expand_bn/moving_mean:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.running_mean""") ) rename_keys.append( (F"""block{b}_expand_bn/moving_variance:0""", F"""encoder.blocks.{hf_b}.expansion.expand_bn.running_var""") ) rename_keys.append( (F"""block{b}_dwconv/depthwise_kernel:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight""") ) rename_keys.append((F"""block{b}_bn/gamma:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight""") ) rename_keys.append((F"""block{b}_bn/beta:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias""") ) rename_keys.append( (F"""block{b}_bn/moving_mean:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean""") ) rename_keys.append( (F"""block{b}_bn/moving_variance:0""", F"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var""") ) rename_keys.append((F"""block{b}_se_reduce/kernel:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.reduce.weight""") ) rename_keys.append((F"""block{b}_se_reduce/bias:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.reduce.bias""") ) rename_keys.append((F"""block{b}_se_expand/kernel:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.expand.weight""") ) rename_keys.append((F"""block{b}_se_expand/bias:0""", F"""encoder.blocks.{hf_b}.squeeze_excite.expand.bias""") ) rename_keys.append( (F"""block{b}_project_conv/kernel:0""", F"""encoder.blocks.{hf_b}.projection.project_conv.weight""") ) rename_keys.append((F"""block{b}_project_bn/gamma:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.weight""") ) rename_keys.append((F"""block{b}_project_bn/beta:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.bias""") ) rename_keys.append( (F"""block{b}_project_bn/moving_mean:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.running_mean""") ) rename_keys.append( (F"""block{b}_project_bn/moving_variance:0""", F"""encoder.blocks.{hf_b}.projection.project_bn.running_var""") ) rename_keys.append(('''top_conv/kernel:0''', '''encoder.top_conv.weight''') ) rename_keys.append(('''top_bn/gamma:0''', '''encoder.top_bn.weight''') ) rename_keys.append(('''top_bn/beta:0''', '''encoder.top_bn.bias''') ) rename_keys.append(('''top_bn/moving_mean:0''', '''encoder.top_bn.running_mean''') ) rename_keys.append(('''top_bn/moving_variance:0''', '''encoder.top_bn.running_var''') ) _lowerCamelCase = {} for item in rename_keys: if item[0] in original_param_names: _lowerCamelCase = '''efficientnet.''' + item[1] _lowerCamelCase = '''classifier.weight''' _lowerCamelCase = '''classifier.bias''' return key_mapping def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : str ) -> Any: for key, value in tf_params.items(): if "normalization" in key: continue _lowerCamelCase = key_mapping[key] if "_conv" in key and "kernel" in key: _lowerCamelCase = torch.from_numpy(lowercase_ ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: _lowerCamelCase = torch.from_numpy(lowercase_ ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: _lowerCamelCase = torch.from_numpy(np.transpose(lowercase_ ) ) else: _lowerCamelCase = torch.from_numpy(lowercase_ ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(lowercase_ ) @torch.no_grad() def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : int , lowercase_ : str , lowercase_ : Union[str, Any] ) -> List[Any]: _lowerCamelCase = model_classes[model_name]( include_top=lowercase_ , weights='''imagenet''' , input_tensor=lowercase_ , input_shape=lowercase_ , pooling=lowercase_ , classes=10_00 , classifier_activation='''softmax''' , ) _lowerCamelCase = original_model.trainable_variables _lowerCamelCase = original_model.non_trainable_variables _lowerCamelCase = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: _lowerCamelCase = param.numpy() _lowerCamelCase = list(tf_params.keys() ) # Load HuggingFace model _lowerCamelCase = get_efficientnet_config(lowercase_ ) _lowerCamelCase = EfficientNetForImageClassification(lowercase_ ).eval() _lowerCamelCase = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print('''Converting parameters...''' ) _lowerCamelCase = rename_keys(lowercase_ ) replace_params(lowercase_ , lowercase_ , lowercase_ ) # Initialize preprocessor and preprocess input image _lowerCamelCase = convert_image_processor(lowercase_ ) _lowerCamelCase = preprocessor(images=prepare_img() , return_tensors='''pt''' ) # HF model inference hf_model.eval() with torch.no_grad(): _lowerCamelCase = hf_model(**lowercase_ ) _lowerCamelCase = outputs.logits.detach().numpy() # Original model inference _lowerCamelCase = False _lowerCamelCase = CONFIG_MAP[model_name]['''image_size'''] _lowerCamelCase = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) _lowerCamelCase = image.img_to_array(lowercase_ ) _lowerCamelCase = np.expand_dims(lowercase_ , axis=0 ) _lowerCamelCase = original_model.predict(lowercase_ ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(lowercase_ , lowercase_ , atol=1e-3 ), "The predicted logits are not the same." print('''Model outputs match!''' ) if save_model: # Create folder to save model if not os.path.isdir(lowercase_ ): os.mkdir(lowercase_ ) # Save converted model and image processor hf_model.save_pretrained(lowercase_ ) preprocessor.save_pretrained(lowercase_ ) if push_to_hub: # Push model and image processor to hub print(F"""Pushing converted {model_name} to the hub...""" ) _lowerCamelCase = F"""efficientnet-{model_name}""" preprocessor.push_to_hub(lowercase_ ) hf_model.push_to_hub(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''b0''', type=str, help='''Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''hf_model''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--save_model''', action='''store_true''', help='''Save model to local''') parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
623
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
1
"""simple docstring""" import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowerCamelCase_( enum.Enum ): '''simple docstring''' lowercase__ : int = 0 lowercase__ : Optional[Any] = 1 lowercase__ : List[str] = 2 @add_end_docstrings(A__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == '''tf''' else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. _lowerCamelCase = None if self.model.config.prefix is not None: _lowerCamelCase = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. _lowerCamelCase = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self._sanitize_parameters(prefix=lowerCamelCase__ , **self._forward_params ) _lowerCamelCase = {**self._preprocess_params, **preprocess_params} _lowerCamelCase = {**self._forward_params, **forward_params} def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ , ): _lowerCamelCase = {} if prefix is not None: _lowerCamelCase = prefix if prefix: _lowerCamelCase = self.tokenizer( lowerCamelCase__ , padding=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=self.framework ) _lowerCamelCase = prefix_inputs['''input_ids'''].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( F"""{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected""" ''' [None, \'hole\']''' ) _lowerCamelCase = handle_long_generation preprocess_params.update(lowerCamelCase__ ) _lowerCamelCase = generate_kwargs _lowerCamelCase = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError('''`return_text` is mutually exclusive with `return_full_text`''' ) if return_tensors is not None: raise ValueError('''`return_full_text` is mutually exclusive with `return_tensors`''' ) _lowerCamelCase = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError('''`return_text` is mutually exclusive with `return_tensors`''' ) _lowerCamelCase = ReturnType.TENSORS if return_type is not None: _lowerCamelCase = return_type if clean_up_tokenization_spaces is not None: _lowerCamelCase = clean_up_tokenization_spaces if stop_sequence is not None: _lowerCamelCase = self.tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) if len(lowerCamelCase__ ) > 1: warnings.warn( '''Stopping on a multiple token sequence is not yet supported on transformers. The first token of''' ''' the stop sequence will be used as the stop sequence string in the interim.''' ) _lowerCamelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({'''add_space_before_punct_symbol''': True} ) return super()._parse_and_tokenize(*lowerCamelCase__ , **lowerCamelCase__ ) def __call__( self , lowerCamelCase__ , **lowerCamelCase__ ): return super().__call__(lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="" , lowerCamelCase__=None , **lowerCamelCase__ ): _lowerCamelCase = self.tokenizer( prefix + prompt_text , padding=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , return_tensors=self.framework ) _lowerCamelCase = prompt_text if handle_long_generation == "hole": _lowerCamelCase = inputs['''input_ids'''].shape[-1] if "max_new_tokens" in generate_kwargs: _lowerCamelCase = generate_kwargs['''max_new_tokens'''] else: _lowerCamelCase = generate_kwargs.get('''max_length''' , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError('''We cannot infer how many new tokens are expected''' ) if cur_len + new_tokens > self.tokenizer.model_max_length: _lowerCamelCase = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( '''We cannot use `hole` to handle this generation the number of desired tokens exceeds the''' ''' models max length''' ) _lowerCamelCase = inputs['''input_ids'''][:, -keep_length:] if "attention_mask" in inputs: _lowerCamelCase = inputs['''attention_mask'''][:, -keep_length:] return inputs def snake_case__ ( self , lowerCamelCase__ , **lowerCamelCase__ ): _lowerCamelCase = model_inputs['''input_ids'''] _lowerCamelCase = model_inputs.get('''attention_mask''' , lowerCamelCase__ ) # Allow empty prompts if input_ids.shape[1] == 0: _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = 1 else: _lowerCamelCase = input_ids.shape[0] _lowerCamelCase = model_inputs.pop('''prompt_text''' ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. _lowerCamelCase = generate_kwargs.pop('''prefix_length''' , 0 ) if prefix_length > 0: _lowerCamelCase = '''max_new_tokens''' in generate_kwargs or ( '''generation_config''' in generate_kwargs and generate_kwargs['''generation_config'''].max_new_tokens is not None ) if not has_max_new_tokens: _lowerCamelCase = generate_kwargs.get('''max_length''' ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length _lowerCamelCase = '''min_new_tokens''' in generate_kwargs or ( '''generation_config''' in generate_kwargs and generate_kwargs['''generation_config'''].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL _lowerCamelCase = self.model.generate(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = generated_sequence.shape[0] if self.framework == "pt": _lowerCamelCase = generated_sequence.reshape(lowerCamelCase__ , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": _lowerCamelCase = tf.reshape(lowerCamelCase__ , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=ReturnType.FULL_TEXT , lowerCamelCase__=True ): _lowerCamelCase = model_outputs['''generated_sequence'''][0] _lowerCamelCase = model_outputs['''input_ids'''] _lowerCamelCase = model_outputs['''prompt_text'''] _lowerCamelCase = generated_sequence.numpy().tolist() _lowerCamelCase = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: _lowerCamelCase = {'''generated_token_ids''': sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text _lowerCamelCase = self.tokenizer.decode( lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: _lowerCamelCase = 0 else: _lowerCamelCase = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ , ) ) if return_type == ReturnType.FULL_TEXT: _lowerCamelCase = prompt_text + text[prompt_length:] else: _lowerCamelCase = text[prompt_length:] _lowerCamelCase = {'''generated_text''': all_text} records.append(lowerCamelCase__ ) return records
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = 'vit_msn' def __init__( self , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-06 , lowerCamelCase__=2_2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=3 , lowerCamelCase__=True , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = qkv_bias
623
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
1
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = ['image_processor', 'tokenizer'] lowercase__ : Tuple = 'AutoImageProcessor' lowercase__ : Optional[int] = 'AutoTokenizer' def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ ): _lowerCamelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , lowerCamelCase__ , ) _lowerCamelCase = kwargs.pop('''feature_extractor''' ) _lowerCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.image_processor _lowerCamelCase = False def __call__( self , *lowerCamelCase__ , **lowerCamelCase__ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''images''' , lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''text''' , lowerCamelCase__ ) if len(lowerCamelCase__ ) > 0: _lowerCamelCase = args[0] _lowerCamelCase = args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: _lowerCamelCase = self.image_processor(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) if text is not None: _lowerCamelCase = self.tokenizer(lowerCamelCase__ , **lowerCamelCase__ ) if text is None: return inputs elif images is None: return encodings else: _lowerCamelCase = encodings['''input_ids'''] return inputs def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) @contextmanager def snake_case__ ( self ): warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) _lowerCamelCase = True _lowerCamelCase = self.tokenizer yield _lowerCamelCase = self.image_processor _lowerCamelCase = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=None ): if added_vocab is None: _lowerCamelCase = self.tokenizer.get_added_vocab() _lowerCamelCase = {} while tokens: _lowerCamelCase = re.search(R'''<s_(.*?)>''' , lowerCamelCase__ , re.IGNORECASE ) if start_token is None: break _lowerCamelCase = start_token.group(1 ) _lowerCamelCase = re.search(RF"""</s_{key}>""" , lowerCamelCase__ , re.IGNORECASE ) _lowerCamelCase = start_token.group() if end_token is None: _lowerCamelCase = tokens.replace(lowerCamelCase__ , '''''' ) else: _lowerCamelCase = end_token.group() _lowerCamelCase = re.escape(lowerCamelCase__ ) _lowerCamelCase = re.escape(lowerCamelCase__ ) _lowerCamelCase = re.search(F"""{start_token_escaped}(.*?){end_token_escaped}""" , lowerCamelCase__ , re.IGNORECASE ) if content is not None: _lowerCamelCase = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _lowerCamelCase = self.tokenajson(lowerCamelCase__ , is_inner_value=lowerCamelCase__ , added_vocab=lowerCamelCase__ ) if value: if len(lowerCamelCase__ ) == 1: _lowerCamelCase = value[0] _lowerCamelCase = value else: # leaf nodes _lowerCamelCase = [] for leaf in content.split(R'''<sep/>''' ): _lowerCamelCase = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _lowerCamelCase = leaf[1:-2] # for categorical special tokens output[key].append(lowerCamelCase__ ) if len(output[key] ) == 1: _lowerCamelCase = output[key][0] _lowerCamelCase = tokens[tokens.find(lowerCamelCase__ ) + len(lowerCamelCase__ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=lowerCamelCase__ , added_vocab=lowerCamelCase__ ) if len(lowerCamelCase__ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def snake_case__ ( self ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , lowerCamelCase__ , ) return self.image_processor_class @property def snake_case__ ( self ): warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , lowerCamelCase__ , ) return self.image_processor
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py __SCREAMING_SNAKE_CASE : Any = '''src/transformers''' __SCREAMING_SNAKE_CASE : Optional[int] = '''docs/source/en/tasks''' def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : List[Any] ) -> Tuple: with open(lowercase_ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: _lowerCamelCase = f.readlines() # Find the start prompt. _lowerCamelCase = 0 while not lines[start_index].startswith(lowercase_ ): start_index += 1 start_index += 1 _lowerCamelCase = start_index while not lines[end_index].startswith(lowercase_ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. __SCREAMING_SNAKE_CASE : Optional[Any] = direct_transformers_import(TRANSFORMERS_PATH) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''asr.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, '''audio_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, '''language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, '''image_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, '''masked_language_modeling.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, '''multiple_choice.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, '''object_detection.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, '''question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, '''semantic_segmentation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, '''sequence_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, '''summarization.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''token_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, '''translation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, '''video_classification.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, '''document_question_answering.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, '''monocular_depth_estimation.md''': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). __SCREAMING_SNAKE_CASE : List[str] = { '''summarization.md''': ('''nllb''',), '''translation.md''': ('''nllb''',), } def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> List[Any]: _lowerCamelCase = TASK_GUIDE_TO_MODELS[task_guide] _lowerCamelCase = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(lowercase_ , set() ) _lowerCamelCase = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"""[{name}](../model_doc/{code})""" for code, name in model_names.items()] ) + "\n" def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Optional[int]=False ) -> str: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = _find_text_in_file( filename=os.path.join(lowercase_ , lowercase_ ) , start_prompt='''<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->''' , end_prompt='''<!--End of the generated tip-->''' , ) _lowerCamelCase = get_model_list_for_task(lowercase_ ) if current_list != new_list: if overwrite: with open(os.path.join(lowercase_ , lowercase_ ) , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"""The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`""" ''' to fix this.''' ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') __SCREAMING_SNAKE_CASE : List[str] = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" from scipy.stats import pearsonr import datasets __SCREAMING_SNAKE_CASE : Dict = ''' Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. ''' __SCREAMING_SNAKE_CASE : int = ''' Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results[\'pearsonr\'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric("pearsonr") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) [\'p-value\', \'pearsonr\'] >>> print(round(results[\'pearsonr\'], 2)) -0.74 >>> print(round(results[\'p-value\'], 2)) 0.15 ''' __SCREAMING_SNAKE_CASE : int = ''' @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): if return_pvalue: _lowerCamelCase = pearsonr(lowerCamelCase__ , lowerCamelCase__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(lowerCamelCase__ , lowerCamelCase__ )[0] )}
623
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __SCREAMING_SNAKE_CASE : Optional[int] = { '''configuration_nezha''': ['''NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''NezhaConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ '''NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''NezhaForNextSentencePrediction''', '''NezhaForMaskedLM''', '''NezhaForPreTraining''', '''NezhaForMultipleChoice''', '''NezhaForQuestionAnswering''', '''NezhaForSequenceClassification''', '''NezhaForTokenClassification''', '''NezhaModel''', '''NezhaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int ) -> int: return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def lowerCAmelCase_( lowercase_ : int ) -> bool: _lowerCamelCase = 0 _lowerCamelCase = number while duplicate > 0: _lowerCamelCase , _lowerCamelCase = divmod(lowercase_ , 10 ) fact_sum += factorial(lowercase_ ) return fact_sum == number if __name__ == "__main__": print('''Program to check whether a number is a Krisnamurthy Number or not.''') __SCREAMING_SNAKE_CASE : Optional[int] = int(input('''Enter number: ''').strip()) print( F"""{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.""" )
623
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int , lowercase_ : int ) -> int: while a != 0: _lowerCamelCase , _lowerCamelCase = b % a, a return b def lowerCAmelCase_( lowercase_ : int , lowercase_ : int ) -> int: if gcd(lowercase_ , lowercase_ ) != 1: _lowerCamelCase = F"""mod inverse of {a!r} and {m!r} does not exist""" raise ValueError(lowercase_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 1, 0, a _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 0, 1, m while va != 0: _lowerCamelCase = ua // va _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = (ua - q * va), (ua - q * va), (ua - q * va), va, va, va return ua % m
623
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
1
"""simple docstring""" import argparse import importlib from pathlib import Path # Test all the extensions added in the setup __SCREAMING_SNAKE_CASE : Optional[int] = [ '''kernels/rwkv/wkv_cuda.cu''', '''kernels/rwkv/wkv_op.cpp''', '''kernels/deformable_detr/ms_deform_attn.h''', '''kernels/deformable_detr/cuda/ms_deform_im2col_cuda.cuh''', '''models/graphormer/algos_graphormer.pyx''', ] def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> Union[str, Any]: # Test all the extensions added in the setup for file in FILES_TO_FIND: if not (transformers_path / file).exists(): return False return True if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[Any] = argparse.ArgumentParser() parser.add_argument('''--check_lib''', action='''store_true''', help='''Whether to check the build or the actual package.''') __SCREAMING_SNAKE_CASE : Any = parser.parse_args() if args.check_lib: __SCREAMING_SNAKE_CASE : str = importlib.import_module('''transformers''') __SCREAMING_SNAKE_CASE : Union[str, Any] = Path(transformers_module.__file__).parent else: __SCREAMING_SNAKE_CASE : str = Path.cwd() / '''build/lib/transformers''' if not test_custom_files_are_present(transformers_path): raise ValueError('''The built release does not contain the custom files. Fix this before going further!''')
623
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0 for i in range(n + 1 )] _lowerCamelCase = 1 _lowerCamelCase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowercase_ ): _lowerCamelCase = 1 _lowerCamelCase = 0 for i in range(lowercase_ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F"""{solution() = }""")
623
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[int] = { '''asapp/sew-d-tiny-100k''': '''https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json''', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : int = 'sew-d' def __init__( self , lowerCamelCase__=3_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__=2 , lowerCamelCase__=5_1_2 , lowerCamelCase__=2_5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=("p2c", "c2p") , lowerCamelCase__="layer_norm" , lowerCamelCase__="gelu_python" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-7 , lowerCamelCase__=1e-5 , lowerCamelCase__="group" , lowerCamelCase__="gelu" , lowerCamelCase__=(6_4, 1_2_8, 1_2_8, 1_2_8, 1_2_8, 2_5_6, 2_5_6, 2_5_6, 2_5_6, 5_1_2, 5_1_2, 5_1_2, 5_1_2) , lowerCamelCase__=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowerCamelCase__=(1_0, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowerCamelCase__=False , lowerCamelCase__=1_2_8 , lowerCamelCase__=1_6 , lowerCamelCase__=True , lowerCamelCase__=0.0_5 , lowerCamelCase__=1_0 , lowerCamelCase__=2 , lowerCamelCase__=0.0 , lowerCamelCase__=1_0 , lowerCamelCase__=0 , lowerCamelCase__="mean" , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=2_5_6 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ , pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) _lowerCamelCase = hidden_size _lowerCamelCase = feat_extract_norm _lowerCamelCase = feat_extract_activation _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = conv_bias _lowerCamelCase = num_conv_pos_embeddings _lowerCamelCase = num_conv_pos_embedding_groups _lowerCamelCase = len(self.conv_dim ) _lowerCamelCase = num_hidden_layers _lowerCamelCase = intermediate_size _lowerCamelCase = squeeze_factor _lowerCamelCase = max_position_embeddings _lowerCamelCase = position_buckets _lowerCamelCase = share_att_key _lowerCamelCase = relative_attention _lowerCamelCase = norm_rel_ebd _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = hidden_act _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_dropout _lowerCamelCase = attention_dropout _lowerCamelCase = activation_dropout _lowerCamelCase = feat_proj_dropout _lowerCamelCase = final_dropout _lowerCamelCase = layer_norm_eps _lowerCamelCase = feature_layer_norm_eps _lowerCamelCase = initializer_range _lowerCamelCase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' F"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" F"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _lowerCamelCase = apply_spec_augment _lowerCamelCase = mask_time_prob _lowerCamelCase = mask_time_length _lowerCamelCase = mask_time_min_masks _lowerCamelCase = mask_feature_prob _lowerCamelCase = mask_feature_length _lowerCamelCase = mask_feature_min_masks # ctc loss _lowerCamelCase = ctc_loss_reduction _lowerCamelCase = ctc_zero_infinity # sequence classification _lowerCamelCase = use_weighted_layer_sum _lowerCamelCase = classifier_proj_size @property def snake_case__ ( self ): return functools.reduce(operator.mul , self.conv_stride , 1 )
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_gpt_neo''': ['''GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoConfig''', '''GPTNeoOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoForCausalLM''', '''GPTNeoForQuestionAnswering''', '''GPTNeoForSequenceClassification''', '''GPTNeoForTokenClassification''', '''GPTNeoModel''', '''GPTNeoPreTrainedModel''', '''load_tf_weights_in_gpt_neo''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''FlaxGPTNeoForCausalLM''', '''FlaxGPTNeoModel''', '''FlaxGPTNeoPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys __SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
1
"""simple docstring""" import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated __SCREAMING_SNAKE_CASE : List[Any] = collections.namedtuple('''_Datasets''', ['''train''', '''validation''', '''test''']) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ __SCREAMING_SNAKE_CASE : List[Any] = '''https://storage.googleapis.com/cvdf-datasets/mnist/''' def lowerCAmelCase_( lowercase_ : Any ) -> List[str]: _lowerCamelCase = numpy.dtype(numpy.uintaa ).newbyteorder('''>''' ) return numpy.frombuffer(bytestream.read(4 ) , dtype=lowercase_ )[0] @deprecated(lowercase_ , '''Please use tf.data to implement this functionality.''' ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> List[str]: print('''Extracting''' , f.name ) with gzip.GzipFile(fileobj=lowercase_ ) as bytestream: _lowerCamelCase = _readaa(lowercase_ ) if magic != 20_51: raise ValueError( '''Invalid magic number %d in MNIST image file: %s''' % (magic, f.name) ) _lowerCamelCase = _readaa(lowercase_ ) _lowerCamelCase = _readaa(lowercase_ ) _lowerCamelCase = _readaa(lowercase_ ) _lowerCamelCase = bytestream.read(rows * cols * num_images ) _lowerCamelCase = numpy.frombuffer(lowercase_ , dtype=numpy.uinta ) _lowerCamelCase = data.reshape(lowercase_ , lowercase_ , lowercase_ , 1 ) return data @deprecated(lowercase_ , '''Please use tf.one_hot on tensors.''' ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : Union[str, Any] ) -> str: _lowerCamelCase = labels_dense.shape[0] _lowerCamelCase = numpy.arange(lowercase_ ) * num_classes _lowerCamelCase = numpy.zeros((num_labels, num_classes) ) _lowerCamelCase = 1 return labels_one_hot @deprecated(lowercase_ , '''Please use tf.data to implement this functionality.''' ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Dict=False , lowercase_ : Union[str, Any]=10 ) -> Optional[Any]: print('''Extracting''' , f.name ) with gzip.GzipFile(fileobj=lowercase_ ) as bytestream: _lowerCamelCase = _readaa(lowercase_ ) if magic != 20_49: raise ValueError( '''Invalid magic number %d in MNIST label file: %s''' % (magic, f.name) ) _lowerCamelCase = _readaa(lowercase_ ) _lowerCamelCase = bytestream.read(lowercase_ ) _lowerCamelCase = numpy.frombuffer(lowercase_ , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(lowercase_ , lowercase_ ) return labels class lowerCamelCase_: '''simple docstring''' @deprecated( lowerCamelCase__ , '''Please use alternatives such as official/mnist/_DataSet.py''' ''' from tensorflow/models.''' , ) def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=dtypes.floataa , lowerCamelCase__=True , lowerCamelCase__=None , ): _lowerCamelCase , _lowerCamelCase = random_seed.get_seed(lowerCamelCase__ ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) _lowerCamelCase = dtypes.as_dtype(lowerCamelCase__ ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError('''Invalid image dtype %r, expected uint8 or float32''' % dtype ) if fake_data: _lowerCamelCase = 1_0_0_0_0 _lowerCamelCase = one_hot else: assert ( images.shape[0] == labels.shape[0] ), F"""images.shape: {images.shape} labels.shape: {labels.shape}""" _lowerCamelCase = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 _lowerCamelCase = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. _lowerCamelCase = images.astype(numpy.floataa ) _lowerCamelCase = numpy.multiply(lowerCamelCase__ , 1.0 / 2_5_5.0 ) _lowerCamelCase = images _lowerCamelCase = labels _lowerCamelCase = 0 _lowerCamelCase = 0 @property def snake_case__ ( self ): return self._images @property def snake_case__ ( self ): return self._labels @property def snake_case__ ( self ): return self._num_examples @property def snake_case__ ( self ): return self._epochs_completed def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=True ): if fake_data: _lowerCamelCase = [1] * 7_8_4 _lowerCamelCase = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(lowerCamelCase__ )], [fake_label for _ in range(lowerCamelCase__ )], ) _lowerCamelCase = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: _lowerCamelCase = numpy.arange(self._num_examples ) numpy.random.shuffle(lowerCamelCase__ ) _lowerCamelCase = self.images[perma] _lowerCamelCase = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch _lowerCamelCase = self._num_examples - start _lowerCamelCase = self._images[start : self._num_examples] _lowerCamelCase = self._labels[start : self._num_examples] # Shuffle the data if shuffle: _lowerCamelCase = numpy.arange(self._num_examples ) numpy.random.shuffle(lowerCamelCase__ ) _lowerCamelCase = self.images[perm] _lowerCamelCase = self.labels[perm] # Start next epoch _lowerCamelCase = 0 _lowerCamelCase = batch_size - rest_num_examples _lowerCamelCase = self._index_in_epoch _lowerCamelCase = self._images[start:end] _lowerCamelCase = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size _lowerCamelCase = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(lowercase_ , '''Please write your own downloading logic.''' ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : List[Any] ) -> Any: if not gfile.Exists(lowercase_ ): gfile.MakeDirs(lowercase_ ) _lowerCamelCase = os.path.join(lowercase_ , lowercase_ ) if not gfile.Exists(lowercase_ ): urllib.request.urlretrieve(lowercase_ , lowercase_ ) # noqa: S310 with gfile.GFile(lowercase_ ) as f: _lowerCamelCase = f.size() print('''Successfully downloaded''' , lowercase_ , lowercase_ , '''bytes.''' ) return filepath @deprecated( lowercase_ , '''Please use alternatives such as:''' ''' tensorflow_datasets.load(\'mnist\')''' ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Union[str, Any]=False , lowercase_ : Dict=False , lowercase_ : Dict=dtypes.floataa , lowercase_ : List[str]=True , lowercase_ : Union[str, Any]=50_00 , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=DEFAULT_SOURCE_URL , ) -> List[Any]: if fake_data: def fake(): return _DataSet( [] , [] , fake_data=lowercase_ , one_hot=lowercase_ , dtype=lowercase_ , seed=lowercase_ ) _lowerCamelCase = fake() _lowerCamelCase = fake() _lowerCamelCase = fake() return _Datasets(train=lowercase_ , validation=lowercase_ , test=lowercase_ ) if not source_url: # empty string check _lowerCamelCase = DEFAULT_SOURCE_URL _lowerCamelCase = '''train-images-idx3-ubyte.gz''' _lowerCamelCase = '''train-labels-idx1-ubyte.gz''' _lowerCamelCase = '''t10k-images-idx3-ubyte.gz''' _lowerCamelCase = '''t10k-labels-idx1-ubyte.gz''' _lowerCamelCase = _maybe_download( lowercase_ , lowercase_ , source_url + train_images_file ) with gfile.Open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = _extract_images(lowercase_ ) _lowerCamelCase = _maybe_download( lowercase_ , lowercase_ , source_url + train_labels_file ) with gfile.Open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = _extract_labels(lowercase_ , one_hot=lowercase_ ) _lowerCamelCase = _maybe_download( lowercase_ , lowercase_ , source_url + test_images_file ) with gfile.Open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = _extract_images(lowercase_ ) _lowerCamelCase = _maybe_download( lowercase_ , lowercase_ , source_url + test_labels_file ) with gfile.Open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = _extract_labels(lowercase_ , one_hot=lowercase_ ) if not 0 <= validation_size <= len(lowercase_ ): _lowerCamelCase = ( '''Validation size should be between 0 and ''' F"""{len(lowercase_ )}. Received: {validation_size}.""" ) raise ValueError(lowercase_ ) _lowerCamelCase = train_images[:validation_size] _lowerCamelCase = train_labels[:validation_size] _lowerCamelCase = train_images[validation_size:] _lowerCamelCase = train_labels[validation_size:] _lowerCamelCase = {'''dtype''': dtype, '''reshape''': reshape, '''seed''': seed} _lowerCamelCase = _DataSet(lowercase_ , lowercase_ , **lowercase_ ) _lowerCamelCase = _DataSet(lowercase_ , lowercase_ , **lowercase_ ) _lowerCamelCase = _DataSet(lowercase_ , lowercase_ , **lowercase_ ) return _Datasets(train=lowercase_ , validation=lowercase_ , test=lowercase_ )
623
"""simple docstring""" from typing import Any class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) _lowerCamelCase = temp.next print() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = Node(lowerCamelCase__ ) _lowerCamelCase = self.head _lowerCamelCase = new_node def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if node_data_a == node_data_a: return else: _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next if node_a is None or node_a is None: return _lowerCamelCase , _lowerCamelCase = node_a.data, node_a.data if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
623
1
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = AudioLDMPipeline lowercase__ : int = TEXT_TO_AUDIO_PARAMS lowercase__ : Dict = TEXT_TO_AUDIO_BATCH_PARAMS lowercase__ : List[str] = frozenset( [ 'num_inference_steps', 'num_waveforms_per_prompt', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=(3_2, 6_4) , class_embed_type='''simple_projection''' , projection_class_embeddings_input_dim=3_2 , class_embeddings_concat=lowerCamelCase__ , ) _lowerCamelCase = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='''scaled_linear''' , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=1 , out_channels=1 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCamelCase = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , projection_dim=3_2 , ) _lowerCamelCase = ClapTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = RobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-random-roberta''' , model_max_length=7_7 ) _lowerCamelCase = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=1_6_0_0_0 , upsample_initial_channel=1_6 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=lowerCamelCase__ , ) _lowerCamelCase = SpeechTaHifiGan(lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''vocoder''': vocoder, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A hammer hitting a wooden surface''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 2_5_6 _lowerCamelCase = audio[:1_0] _lowerCamelCase = np.array( [-0.0_0_5_0, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_3, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_3] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * [inputs['''prompt''']] # forward _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.audios[0] _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] _lowerCamelCase = audioldm_pipe.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCamelCase__ , return_tensors='''pt''' , ) _lowerCamelCase = text_inputs['''input_ids'''].to(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.text_encoder( lowerCamelCase__ , ) _lowerCamelCase = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state _lowerCamelCase = F.normalize(lowerCamelCase__ , dim=-1 ) _lowerCamelCase = prompt_embeds # forward _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] # forward _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.audios[0] _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] _lowerCamelCase = [] for p in [prompt, negative_prompt]: _lowerCamelCase = audioldm_pipe.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=lowerCamelCase__ , return_tensors='''pt''' , ) _lowerCamelCase = text_inputs['''input_ids'''].to(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.text_encoder( lowerCamelCase__ , ) _lowerCamelCase = text_embeds.text_embeds # additional L_2 normalization over each hidden-state _lowerCamelCase = F.normalize(lowerCamelCase__ , dim=-1 ) embeds.append(lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = embeds # forward _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = '''egg cracking''' _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 2_5_6 _lowerCamelCase = audio[:1_0] _lowerCamelCase = np.array( [-0.0_0_5_1, 0.0_0_5_0, -0.0_0_6_0, 0.0_0_3_4, -0.0_0_2_6, 0.0_0_3_3, -0.0_0_2_7, 0.0_0_3_3, -0.0_0_2_8, 0.0_0_3_2] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A hammer hitting a wooden surface''' # test num_waveforms_per_prompt=1 (default) _lowerCamelCase = audioldm_pipe(lowerCamelCase__ , num_inference_steps=2 ).audios assert audios.shape == (1, 2_5_6) # test num_waveforms_per_prompt=1 (default) for batch of prompts _lowerCamelCase = 2 _lowerCamelCase = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 2_5_6) # test num_waveforms_per_prompt for single prompt _lowerCamelCase = 2 _lowerCamelCase = audioldm_pipe(lowerCamelCase__ , num_inference_steps=2 , num_waveforms_per_prompt=lowerCamelCase__ ).audios assert audios.shape == (num_waveforms_per_prompt, 2_5_6) # test num_waveforms_per_prompt for batch of prompts _lowerCamelCase = 2 _lowerCamelCase = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=lowerCamelCase__ ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 2_5_6) def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.vocoder.config.sampling_rate _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe(audio_length_in_s=0.0_1_6 , **lowerCamelCase__ ) _lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) / vocoder_sampling_rate == 0.0_1_6 _lowerCamelCase = audioldm_pipe(audio_length_in_s=0.0_3_2 , **lowerCamelCase__ ) _lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) / vocoder_sampling_rate == 0.0_3_2 def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = AudioLDMPipeline(**lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = ['''hey'''] _lowerCamelCase = audioldm_pipe(lowerCamelCase__ , num_inference_steps=1 ) _lowerCamelCase = output.audios.shape assert audio_shape == (1, 2_5_6) _lowerCamelCase = audioldm_pipe.vocoder.config config.model_in_dim *= 2 _lowerCamelCase = SpeechTaHifiGan(lowerCamelCase__ ).to(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe(lowerCamelCase__ , num_inference_steps=1 ) _lowerCamelCase = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 2_5_6) def snake_case__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCamelCase__ ) def snake_case__ ( self ): self._test_inference_batch_single_identical(test_mean_pixel_difference=lowerCamelCase__ ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def snake_case__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCamelCase__ ) @slow class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 8, 1_2_8, 1_6) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A hammer hitting a wooden surface''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 2.5, } return inputs def snake_case__ ( self ): _lowerCamelCase = AudioLDMPipeline.from_pretrained('''cvssp/audioldm''' ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = 2_5 _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ).audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 8_1_9_2_0 _lowerCamelCase = audio[7_7_2_3_0:7_7_2_4_0] _lowerCamelCase = np.array( [-0.4_8_8_4, -0.4_6_0_7, 0.0_0_2_3, 0.5_0_0_7, 0.5_8_9_6, 0.5_1_5_1, 0.3_8_1_3, -0.0_2_0_8, -0.3_6_8_7, -0.4_3_1_5] ) _lowerCamelCase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def snake_case__ ( self ): _lowerCamelCase = AudioLDMPipeline.from_pretrained('''cvssp/audioldm''' ) _lowerCamelCase = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) _lowerCamelCase = audioldm_pipe.to(lowerCamelCase__ ) audioldm_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = audioldm_pipe(**lowerCamelCase__ ).audios[0] assert audio.ndim == 1 assert len(lowerCamelCase__ ) == 8_1_9_2_0 _lowerCamelCase = audio[2_7_7_8_0:2_7_7_9_0] _lowerCamelCase = np.array([-0.2_1_3_1, -0.0_8_7_3, -0.0_1_2_4, -0.0_1_8_9, 0.0_5_6_9, 0.1_3_7_3, 0.1_8_8_3, 0.2_8_8_6, 0.3_2_9_7, 0.2_2_1_2] ) _lowerCamelCase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
623
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __SCREAMING_SNAKE_CASE : Optional[Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[Any]: from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> List[str]: from diffusers.utils.testing_utils import pytest_terminal_summary_main _lowerCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
623
1
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : float ) -> np.ndarray: return np.where(vector > 0 , lowercase_ , (alpha * (np.exp(lowercase_ ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
1
"""simple docstring""" import os import sys import unittest __SCREAMING_SNAKE_CASE : int = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) __SCREAMING_SNAKE_CASE : List[str] = os.path.join('''tests''', '''models''', '''bert''', '''test_modeling_bert.py''') __SCREAMING_SNAKE_CASE : Union[str, Any] = os.path.join('''tests''', '''models''', '''blip''', '''test_modeling_blip.py''') class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = get_test_to_tester_mapping(lowerCamelCase__ ) _lowerCamelCase = get_test_to_tester_mapping(lowerCamelCase__ ) _lowerCamelCase = {'''BertModelTest''': '''BertModelTester'''} _lowerCamelCase = { '''BlipModelTest''': '''BlipModelTester''', '''BlipTextImageModelTest''': '''BlipTextImageModelsModelTester''', '''BlipTextModelTest''': '''BlipTextModelTester''', '''BlipTextRetrievalModelTest''': '''BlipTextRetrievalModelTester''', '''BlipVQAModelTest''': '''BlipVQAModelTester''', '''BlipVisionModelTest''': '''BlipVisionModelTester''', } self.assertEqual(get_test_info.to_json(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(get_test_info.to_json(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = get_model_to_test_mapping(lowerCamelCase__ ) _lowerCamelCase = get_model_to_test_mapping(lowerCamelCase__ ) _lowerCamelCase = { '''BertForMaskedLM''': ['''BertModelTest'''], '''BertForMultipleChoice''': ['''BertModelTest'''], '''BertForNextSentencePrediction''': ['''BertModelTest'''], '''BertForPreTraining''': ['''BertModelTest'''], '''BertForQuestionAnswering''': ['''BertModelTest'''], '''BertForSequenceClassification''': ['''BertModelTest'''], '''BertForTokenClassification''': ['''BertModelTest'''], '''BertLMHeadModel''': ['''BertModelTest'''], '''BertModel''': ['''BertModelTest'''], } _lowerCamelCase = { '''BlipForConditionalGeneration''': ['''BlipTextImageModelTest'''], '''BlipForImageTextRetrieval''': ['''BlipTextRetrievalModelTest'''], '''BlipForQuestionAnswering''': ['''BlipVQAModelTest'''], '''BlipModel''': ['''BlipModelTest'''], '''BlipTextModel''': ['''BlipTextModelTest'''], '''BlipVisionModel''': ['''BlipVisionModelTest'''], } self.assertEqual(get_test_info.to_json(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(get_test_info.to_json(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = get_model_to_tester_mapping(lowerCamelCase__ ) _lowerCamelCase = get_model_to_tester_mapping(lowerCamelCase__ ) _lowerCamelCase = { '''BertForMaskedLM''': ['''BertModelTester'''], '''BertForMultipleChoice''': ['''BertModelTester'''], '''BertForNextSentencePrediction''': ['''BertModelTester'''], '''BertForPreTraining''': ['''BertModelTester'''], '''BertForQuestionAnswering''': ['''BertModelTester'''], '''BertForSequenceClassification''': ['''BertModelTester'''], '''BertForTokenClassification''': ['''BertModelTester'''], '''BertLMHeadModel''': ['''BertModelTester'''], '''BertModel''': ['''BertModelTester'''], } _lowerCamelCase = { '''BlipForConditionalGeneration''': ['''BlipTextImageModelsModelTester'''], '''BlipForImageTextRetrieval''': ['''BlipTextRetrievalModelTester'''], '''BlipForQuestionAnswering''': ['''BlipVQAModelTester'''], '''BlipModel''': ['''BlipModelTester'''], '''BlipTextModel''': ['''BlipTextModelTester'''], '''BlipVisionModel''': ['''BlipVisionModelTester'''], } self.assertEqual(get_test_info.to_json(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(get_test_info.to_json(lowerCamelCase__ ) , lowerCamelCase__ )
623
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : Dict = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} __SCREAMING_SNAKE_CASE : str = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def lowerCAmelCase_( lowercase_ : dict[int, list[int]] , lowercase_ : int , lowercase_ : list[bool] ) -> list[int]: _lowerCamelCase = True _lowerCamelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(lowercase_ , lowercase_ , lowercase_ ) order.append(lowercase_ ) return order def lowerCAmelCase_( lowercase_ : dict[int, list[int]] , lowercase_ : int , lowercase_ : list[bool] ) -> list[int]: _lowerCamelCase = True _lowerCamelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(lowercase_ , lowercase_ , lowercase_ ) return component def lowerCAmelCase_( lowercase_ : dict[int, list[int]] ) -> list[list[int]]: _lowerCamelCase = len(lowercase_ ) * [False] _lowerCamelCase = {vert: [] for vert in range(len(lowercase_ ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(lowercase_ ) _lowerCamelCase = [] for i, was_visited in enumerate(lowercase_ ): if not was_visited: order += topology_sort(lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = [] _lowerCamelCase = len(lowercase_ ) * [False] for i in range(len(lowercase_ ) ): _lowerCamelCase = order[len(lowercase_ ) - i - 1] if not visited[vert]: _lowerCamelCase = find_components(lowercase_ , lowercase_ , lowercase_ ) components_list.append(lowercase_ ) return components_list
623
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
1
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml __SCREAMING_SNAKE_CASE : Optional[int] = NewType('''DataClass''', Any) __SCREAMING_SNAKE_CASE : Dict = NewType('''DataClassType''', Any) def lowerCAmelCase_( lowercase_ : int ) -> int: if isinstance(lowercase_ , lowercase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def lowerCAmelCase_( lowercase_ : list ) -> Callable[[str], Any]: _lowerCamelCase = {str(lowercase_ ): choice for choice in choices} return lambda lowercase_ : str_to_choice.get(lowercase_ , lowercase_ ) def lowerCAmelCase_( *, lowercase_ : Union[str, List[str]] = None , lowercase_ : str = None , lowercase_ : Any = dataclasses.MISSING , lowercase_ : Callable[[], Any] = dataclasses.MISSING , lowercase_ : dict = None , **lowercase_ : int , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls _lowerCamelCase = {} if aliases is not None: _lowerCamelCase = aliases if help is not None: _lowerCamelCase = help return dataclasses.field(metadata=lowercase_ , default=lowercase_ , default_factory=lowercase_ , **lowercase_ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Iterable[DataClassType] def __init__( self , lowerCamelCase__ , **lowerCamelCase__ ): # To make the default appear when using --help if "formatter_class" not in kwargs: _lowerCamelCase = ArgumentDefaultsHelpFormatter super().__init__(**lowerCamelCase__ ) if dataclasses.is_dataclass(lowerCamelCase__ ): _lowerCamelCase = [dataclass_types] _lowerCamelCase = list(lowerCamelCase__ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(lowerCamelCase__ ) @staticmethod def snake_case__ ( lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = F"""--{field.name}""" _lowerCamelCase = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , lowerCamelCase__ ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) _lowerCamelCase = kwargs.pop('''aliases''' , [] ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = [aliases] _lowerCamelCase = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(lowerCamelCase__ , '''UnionType''' ) and isinstance(lowerCamelCase__ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(lowerCamelCase__ ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' F""" Problem encountered in field '{field.name}'.""" ) if type(lowerCamelCase__ ) not in field.type.__args__: # filter `str` in Union _lowerCamelCase = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] _lowerCamelCase = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) _lowerCamelCase = ( field.type.__args__[0] if isinstance(lowerCamelCase__ , field.type.__args__[1] ) else field.type.__args__[1] ) _lowerCamelCase = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) _lowerCamelCase = {} if origin_type is Literal or (isinstance(field.type , lowerCamelCase__ ) and issubclass(field.type , lowerCamelCase__ )): if origin_type is Literal: _lowerCamelCase = field.type.__args__ else: _lowerCamelCase = [x.value for x in field.type] _lowerCamelCase = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: _lowerCamelCase = field.default else: _lowerCamelCase = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument _lowerCamelCase = copy(lowerCamelCase__ ) # Hack because type=bool in argparse does not behave as we want. _lowerCamelCase = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. _lowerCamelCase = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way _lowerCamelCase = default # This tells argparse we accept 0 or 1 value after --field_name _lowerCamelCase = '''?''' # This is the value that will get picked if we do --field_name (without value) _lowerCamelCase = True elif isclass(lowerCamelCase__ ) and issubclass(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = field.type.__args__[0] _lowerCamelCase = '''+''' if field.default_factory is not dataclasses.MISSING: _lowerCamelCase = field.default_factory() elif field.default is dataclasses.MISSING: _lowerCamelCase = True else: _lowerCamelCase = field.type if field.default is not dataclasses.MISSING: _lowerCamelCase = field.default elif field.default_factory is not dataclasses.MISSING: _lowerCamelCase = field.default_factory() else: _lowerCamelCase = True parser.add_argument(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): _lowerCamelCase = False parser.add_argument(F"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): if hasattr(lowerCamelCase__ , '''_argument_group_name''' ): _lowerCamelCase = self.add_argument_group(dtype._argument_group_name ) else: _lowerCamelCase = self try: _lowerCamelCase = get_type_hints(lowerCamelCase__ ) except NameError: raise RuntimeError( F"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(lowerCamelCase__ ): _lowerCamelCase = '''.'''.join(map(lowerCamelCase__ , sys.version_info[:3] ) ) raise RuntimeError( F"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(lowerCamelCase__ ): if not field.init: continue _lowerCamelCase = type_hints[field.name] self._parse_dataclass_field(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=None , ): if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): _lowerCamelCase = [] if args_filename: args_files.append(Path(lowerCamelCase__ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values _lowerCamelCase = ArgumentParser() args_file_parser.add_argument(lowerCamelCase__ , type=lowerCamelCase__ , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) _lowerCamelCase , _lowerCamelCase = args_file_parser.parse_known_args(args=lowerCamelCase__ ) _lowerCamelCase = vars(lowerCamelCase__ ).get(args_file_flag.lstrip('''-''' ) , lowerCamelCase__ ) if cmd_args_file_paths: args_files.extend([Path(lowerCamelCase__ ) for p in cmd_args_file_paths] ) _lowerCamelCase = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last _lowerCamelCase = file_args + args if args is not None else file_args + sys.argv[1:] _lowerCamelCase , _lowerCamelCase = self.parse_known_args(args=lowerCamelCase__ ) _lowerCamelCase = [] for dtype in self.dataclass_types: _lowerCamelCase = {f.name for f in dataclasses.fields(lowerCamelCase__ ) if f.init} _lowerCamelCase = {k: v for k, v in vars(lowerCamelCase__ ).items() if k in keys} for k in keys: delattr(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = dtype(**lowerCamelCase__ ) outputs.append(lowerCamelCase__ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(lowerCamelCase__ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): _lowerCamelCase = set(args.keys() ) _lowerCamelCase = [] for dtype in self.dataclass_types: _lowerCamelCase = {f.name for f in dataclasses.fields(lowerCamelCase__ ) if f.init} _lowerCamelCase = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) _lowerCamelCase = dtype(**lowerCamelCase__ ) outputs.append(lowerCamelCase__ ) if not allow_extra_keys and unused_keys: raise ValueError(F"""Some keys are not used by the HfArgumentParser: {sorted(lowerCamelCase__ )}""" ) return tuple(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): with open(Path(lowerCamelCase__ ) , encoding='''utf-8''' ) as open_json_file: _lowerCamelCase = json.loads(open_json_file.read() ) _lowerCamelCase = self.parse_dict(lowerCamelCase__ , allow_extra_keys=lowerCamelCase__ ) return tuple(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): _lowerCamelCase = self.parse_dict(yaml.safe_load(Path(lowerCamelCase__ ).read_text() ) , allow_extra_keys=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
623
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
1
"""simple docstring""" from __future__ import annotations from random import random class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ = None ): _lowerCamelCase = value _lowerCamelCase = random() _lowerCamelCase = None _lowerCamelCase = None def __repr__( self ): from pprint import pformat if self.left is None and self.right is None: return F"""'{self.value}: {self.prior:.5}'""" else: return pformat( {F"""{self.value}: {self.prior:.5}""": (self.left, self.right)} , indent=1 ) def __str__( self ): _lowerCamelCase = str(self.value ) + ''' ''' _lowerCamelCase = str(self.left or '''''' ) _lowerCamelCase = str(self.right or '''''' ) return value + left + right def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : int ) -> tuple[Node | None, Node | None]: if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: _lowerCamelCase , _lowerCamelCase = split(root.left , lowercase_ ) return left, root else: _lowerCamelCase , _lowerCamelCase = split(root.right , lowercase_ ) return root, right def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : Node | None ) -> Node | None: if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: _lowerCamelCase = merge(left.right , lowercase_ ) return left else: _lowerCamelCase = merge(lowercase_ , right.left ) return right def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : int ) -> Node | None: _lowerCamelCase = Node(lowercase_ ) _lowerCamelCase , _lowerCamelCase = split(lowercase_ , lowercase_ ) return merge(merge(lowercase_ , lowercase_ ) , lowercase_ ) def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : int ) -> Node | None: _lowerCamelCase , _lowerCamelCase = split(lowercase_ , value - 1 ) _lowerCamelCase , _lowerCamelCase = split(lowercase_ , lowercase_ ) return merge(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : Node | None ) -> None: if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def lowerCAmelCase_( lowercase_ : Node | None , lowercase_ : str ) -> Node | None: for arg in args.split(): if arg[0] == "+": _lowerCamelCase = insert(lowercase_ , int(arg[1:] ) ) elif arg[0] == "-": _lowerCamelCase = erase(lowercase_ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def lowerCAmelCase_( ) -> None: _lowerCamelCase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) _lowerCamelCase = input() while args != "q": _lowerCamelCase = interact_treap(lowercase_ , lowercase_ ) print(lowercase_ ) _lowerCamelCase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
623
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowerCamelCase_( metaclass=A__ ): '''simple docstring''' lowercase__ : Optional[int] = ['transformers', 'torch', 'note_seq'] def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def snake_case__ ( cls , *lowerCamelCase__ , **lowerCamelCase__ ): requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def snake_case__ ( cls , *lowerCamelCase__ , **lowerCamelCase__ ): requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
1
"""simple docstring""" from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time __SCREAMING_SNAKE_CASE : Union[str, Any] = Lock() def lowerCAmelCase_( lowercase_ : Any , lowercase_ : int , lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : Tuple , lowercase_ : str , lowercase_ : int ) -> Dict: global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(lowercase_ ) process_lock.release() # receive your right neighbor's value process_lock.acquire() _lowerCamelCase = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left _lowerCamelCase = min(lowercase_ , lowercase_ ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(lowercase_ ) process_lock.release() # receive your left neighbor's value process_lock.acquire() _lowerCamelCase = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right _lowerCamelCase = max(lowercase_ , lowercase_ ) # after all swaps are performed, send the values back to main result_pipe[1].send(lowercase_ ) def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> Tuple: _lowerCamelCase = [] _lowerCamelCase = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop _lowerCamelCase = Pipe() _lowerCamelCase = Pipe() process_array_.append( Process( target=lowercase_ , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) _lowerCamelCase = temp_rs _lowerCamelCase = temp_rr for i in range(1 , len(lowercase_ ) - 1 ): _lowerCamelCase = Pipe() _lowerCamelCase = Pipe() process_array_.append( Process( target=lowercase_ , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) _lowerCamelCase = temp_rs _lowerCamelCase = temp_rr process_array_.append( Process( target=lowercase_ , args=( len(lowercase_ ) - 1, arr[len(lowercase_ ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(lowercase_ ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(lowercase_ ) ): _lowerCamelCase = result_pipe[p][0].recv() process_array_[p].join() return arr def lowerCAmelCase_( ) -> str: _lowerCamelCase = list(range(10 , 0 , -1 ) ) print('''Initial List''' ) print(*lowercase_ ) _lowerCamelCase = odd_even_transposition(lowercase_ ) print('''Sorted List\n''' ) print(*lowercase_ ) if __name__ == "__main__": main()
623
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import mpmath # for roots of unity import numpy as np class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None ): # Input as list _lowerCamelCase = list(poly_a or [0] )[:] _lowerCamelCase = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() _lowerCamelCase = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() _lowerCamelCase = len(self.polyB ) # Add 0 to make lengths equal a power of 2 _lowerCamelCase = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform _lowerCamelCase = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product _lowerCamelCase = self.__multiply() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [[x] for x in self.polyA] if which == '''A''' else [[x] for x in self.polyB] # Corner case if len(lowerCamelCase__ ) <= 1: return dft[0] # _lowerCamelCase = self.c_max_length // 2 while next_ncol > 0: _lowerCamelCase = [[] for i in range(lowerCamelCase__ )] _lowerCamelCase = self.root**next_ncol # First half of next step _lowerCamelCase = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(lowerCamelCase__ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step _lowerCamelCase = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(lowerCamelCase__ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update _lowerCamelCase = new_dft _lowerCamelCase = next_ncol // 2 return dft[0] def snake_case__ ( self ): _lowerCamelCase = self.__dft('''A''' ) _lowerCamelCase = self.__dft('''B''' ) _lowerCamelCase = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT _lowerCamelCase = 2 while next_ncol <= self.c_max_length: _lowerCamelCase = [[] for i in range(lowerCamelCase__ )] _lowerCamelCase = self.root ** (next_ncol // 2) _lowerCamelCase = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update _lowerCamelCase = new_inverse_c next_ncol *= 2 # Unpack _lowerCamelCase = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self ): _lowerCamelCase = '''A = ''' + ''' + '''.join( F"""{coef}*x^{i}""" for coef, i in enumerate(self.polyA[: self.len_A] ) ) _lowerCamelCase = '''B = ''' + ''' + '''.join( F"""{coef}*x^{i}""" for coef, i in enumerate(self.polyB[: self.len_B] ) ) _lowerCamelCase = '''A*B = ''' + ''' + '''.join( F"""{coef}*x^{i}""" for coef, i in enumerate(self.product ) ) return F"""{a}\n{b}\n{c}""" # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope _lowerCamelCase = self.vocab_size - 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) _lowerCamelCase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = OpenAIGPTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , head_mask=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = OpenAIGPTLMHeadModel(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = OpenAIGPTDoubleHeadsModel(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = OpenAIGPTForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) lowercase__ : List[Any] = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly lowercase__ : Union[str, Any] = ( { 'feature-extraction': OpenAIGPTModel, 'text-classification': OpenAIGPTForSequenceClassification, 'text-generation': OpenAIGPTLMHeadModel, 'zero-shot': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = super()._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": _lowerCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=lowerCamelCase__ , ) _lowerCamelCase = inputs_dict['''labels'''] _lowerCamelCase = inputs_dict['''labels'''] _lowerCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=lowerCamelCase__ , ) _lowerCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase__ ) return inputs_dict def snake_case__ ( self ): _lowerCamelCase = OpenAIGPTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , n_embd=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = OpenAIGPTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = OpenAIGPTLMHeadModel.from_pretrained('''openai-gpt''' ) model.to(lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]] , dtype=torch.long , device=lowerCamelCase__ ) # the president is _lowerCamelCase = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the _lowerCamelCase = model.generate(lowerCamelCase__ , do_sample=lowerCamelCase__ ) self.assertListEqual(output_ids[0].tolist() , lowerCamelCase__ )
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Tuple = { '''microsoft/git-base''': '''https://huggingface.co/microsoft/git-base/resolve/main/config.json''', } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Dict = 'git_vision_model' def __init__( self , lowerCamelCase__=7_6_8 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3 , lowerCamelCase__=2_2_4 , lowerCamelCase__=1_6 , lowerCamelCase__="quick_gelu" , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = hidden_size _lowerCamelCase = intermediate_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = num_channels _lowerCamelCase = patch_size _lowerCamelCase = image_size _lowerCamelCase = initializer_range _lowerCamelCase = attention_dropout _lowerCamelCase = layer_norm_eps _lowerCamelCase = hidden_act @classmethod def snake_case__ ( cls , lowerCamelCase__ , **lowerCamelCase__ ): cls._set_token_in_kwargs(lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get('''model_type''' ) == "git": _lowerCamelCase = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = 'git' def __init__( self , lowerCamelCase__=None , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=6 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=0 , lowerCamelCase__="absolute" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=1_0_1 , lowerCamelCase__=1_0_2 , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__(bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , pad_token_id=lowerCamelCase__ , **lowerCamelCase__ ) if vision_config is None: _lowerCamelCase = {} logger.info('''vision_config is None. initializing the GitVisionConfig with default values.''' ) _lowerCamelCase = GitVisionConfig(**lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = position_embedding_type _lowerCamelCase = use_cache _lowerCamelCase = tie_word_embeddings _lowerCamelCase = num_image_with_embedding _lowerCamelCase = bos_token_id _lowerCamelCase = eos_token_id def snake_case__ ( self ): _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.vision_config.to_dict() _lowerCamelCase = self.__class__.model_type return output
623
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_input_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self ): return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BioGptModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = BioGptForCausalLM(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = BioGptModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() # create attention mask _lowerCamelCase = torch.ones(input_ids.shape , dtype=torch.long , device=lowerCamelCase__ ) _lowerCamelCase = self.seq_length // 2 _lowerCamelCase = 0 # first forward pass _lowerCamelCase , _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ ).to_tuple() # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids _lowerCamelCase = ids_tensor((1,) , lowerCamelCase__ ).item() + 1 _lowerCamelCase = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) _lowerCamelCase = random_other_next_tokens # append to next input_ids and attn_mask _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=lowerCamelCase__ )] , dim=1 , ) # get two different outputs _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ , attention_mask=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, -1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = BioGptModel(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = torch.ones(input_ids.shape , dtype=torch.long , device=lowerCamelCase__ ) # first forward pass _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids _lowerCamelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCamelCase = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , past_key_values=lowerCamelCase__ )[ '''last_hidden_state''' ] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = BioGptForCausalLM(lowerCamelCase__ ) model.to(lowerCamelCase__ ) if gradient_checkpointing: model.gradient_checkpointing_enable() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def snake_case__ ( self , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = BioGptModel(lowerCamelCase__ ) _lowerCamelCase = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_0_1 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.0_1 ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , *lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = BioGptForTokenClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) lowercase__ : List[str] = (BioGptForCausalLM,) if is_torch_available() else () lowercase__ : Union[str, Any] = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) lowercase__ : Dict = False def snake_case__ ( self ): _lowerCamelCase = BioGptModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCamelCase = type self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*lowerCamelCase__ , gradient_checkpointing=lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''' ) model.to(lowerCamelCase__ ) _lowerCamelCase = BioGptTokenizer.from_pretrained('''microsoft/biogpt''' ) _lowerCamelCase = '''left''' # Define PAD Token = EOS Token = 50256 _lowerCamelCase = tokenizer.eos_token _lowerCamelCase = model.config.eos_token_id # use different length sentences to test batching _lowerCamelCase = [ '''Hello, my dog is a little''', '''Today, I''', ] _lowerCamelCase = tokenizer(lowerCamelCase__ , return_tensors='''pt''' , padding=lowerCamelCase__ ) _lowerCamelCase = inputs['''input_ids'''].to(lowerCamelCase__ ) _lowerCamelCase = model.generate( input_ids=lowerCamelCase__ , attention_mask=inputs['''attention_mask'''].to(lowerCamelCase__ ) , ) _lowerCamelCase = tokenizer(sentences[0] , return_tensors='''pt''' ).input_ids.to(lowerCamelCase__ ) _lowerCamelCase = model.generate(input_ids=lowerCamelCase__ ) _lowerCamelCase = inputs_non_padded.shape[-1] - inputs['''attention_mask'''][-1].long().sum().cpu().item() _lowerCamelCase = tokenizer(sentences[1] , return_tensors='''pt''' ).input_ids.to(lowerCamelCase__ ) _lowerCamelCase = model.generate(input_ids=lowerCamelCase__ , max_length=model.config.max_length - num_paddings ) _lowerCamelCase = tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCamelCase__ ) _lowerCamelCase = [ '''Hello, my dog is a little bit bigger than a little bit.''', '''Today, I have a good idea of how to use the information''', ] self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , [non_padded_sentence, padded_sentence] ) @slow def snake_case__ ( self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = BioGptModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = 3 _lowerCamelCase = input_dict['''input_ids'''] _lowerCamelCase = input_ids.ne(1 ).to(lowerCamelCase__ ) _lowerCamelCase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCamelCase = BioGptForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = 3 _lowerCamelCase = '''multi_label_classification''' _lowerCamelCase = input_dict['''input_ids'''] _lowerCamelCase = input_ids.ne(1 ).to(lowerCamelCase__ ) _lowerCamelCase = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _lowerCamelCase = BioGptForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''' ) _lowerCamelCase = torch.tensor([[2, 4_8_0_5, 9, 6_5_6, 2_1]] ) _lowerCamelCase = model(lowerCamelCase__ )[0] _lowerCamelCase = 4_2_3_8_4 _lowerCamelCase = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor( [[[-9.5_2_3_6, -9.8_9_1_8, 1_0.4_5_5_7], [-1_1.0_4_6_9, -9.6_4_2_3, 8.1_0_2_2], [-8.8_6_6_4, -7.8_8_2_6, 5.5_3_2_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase__ , atol=1e-4 ) ) @slow def snake_case__ ( self ): _lowerCamelCase = BioGptTokenizer.from_pretrained('''microsoft/biogpt''' ) _lowerCamelCase = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''' ) model.to(lowerCamelCase__ ) torch.manual_seed(0 ) _lowerCamelCase = tokenizer('''COVID-19 is''' , return_tensors='''pt''' ).to(lowerCamelCase__ ) _lowerCamelCase = model.generate( **lowerCamelCase__ , min_length=1_0_0 , max_length=1_0_2_4 , num_beams=5 , early_stopping=lowerCamelCase__ , ) _lowerCamelCase = tokenizer.decode(output_ids[0] , skip_special_tokens=lowerCamelCase__ ) _lowerCamelCase = ( '''COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the''' ''' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and''' ''' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),''' ''' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and''' ''' more than 800,000 deaths.''' ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ )
623
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
1
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import warnings from .generation import TFGenerationMixin class lowerCamelCase_( A__ ): '''simple docstring''' warnings.warn( 'Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ' 'be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.', A__, )
623
1
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def lowerCAmelCase_( lowercase_ : int ) -> str: if not isinstance(lowercase_ , lowercase_ ): raise TypeError('''Undefined for non-integers''' ) elif precision < 1: raise ValueError('''Undefined for non-natural numbers''' ) _lowerCamelCase = precision _lowerCamelCase = ceil(precision / 14 ) _lowerCamelCase = 42_68_80 * Decimal(1_00_05 ).sqrt() _lowerCamelCase = 1 _lowerCamelCase = 13_59_14_09 _lowerCamelCase = Decimal(lowercase_ ) for k in range(1 , lowercase_ ): _lowerCamelCase = factorial(6 * k ) // (factorial(3 * k ) * factorial(lowercase_ ) ** 3) linear_term += 5_45_14_01_34 exponential_term *= -26_25_37_41_26_40_76_80_00 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = 5_0 print(F"""The first {n} digits of pi is: {pi(n)}""")
623
"""simple docstring""" import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: # Load configuration defined in the metadata file with open(lowercase_ ) as metadata_file: _lowerCamelCase = json.load(lowercase_ ) _lowerCamelCase = LukeConfig(use_entity_aware_attention=lowercase_ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) # Load the entity vocab file _lowerCamelCase = load_entity_vocab(lowercase_ ) _lowerCamelCase = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks _lowerCamelCase = AddedToken('''<ent>''' , lstrip=lowercase_ , rstrip=lowercase_ ) _lowerCamelCase = AddedToken('''<ent2>''' , lstrip=lowercase_ , rstrip=lowercase_ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(lowercase_ ) with open(os.path.join(lowercase_ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ ) # Initialize the embeddings of the special tokens _lowerCamelCase = state_dict['''embeddings.word_embeddings.weight'''] _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) _lowerCamelCase = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) _lowerCamelCase = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: _lowerCamelCase = F"""encoder.layer.{layer_index}.attention.self.""" _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] _lowerCamelCase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks _lowerCamelCase = state_dict['''entity_embeddings.entity_embeddings.weight'''] _lowerCamelCase = entity_emb[entity_vocab['''[MASK]''']] _lowerCamelCase = LukeModel(config=lowercase_ ).eval() _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) if not (len(lowercase_ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {", ".join(lowercase_ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {", ".join([key for key in unexpected_keys if not (key.startswith("entity_predictions" ) or key.startswith("lm_head" ))] )}""" ) # Check outputs _lowerCamelCase = LukeTokenizer.from_pretrained(lowercase_ , task='''entity_classification''' ) _lowerCamelCase = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) _lowerCamelCase = (39, 42) _lowerCamelCase = tokenizer(lowercase_ , entity_spans=[span] , add_prefix_space=lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) # Verify word hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 42, 10_24) ) _lowerCamelCase = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base _lowerCamelCase = torch.Size((1, 42, 7_68) ) _lowerCamelCase = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": _lowerCamelCase = torch.Size((1, 1, 10_24) ) _lowerCamelCase = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base _lowerCamelCase = torch.Size((1, 1, 7_68) ) _lowerCamelCase = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowercase_ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(lowercase_ ) ) model.save_pretrained(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Any: _lowerCamelCase = {} with open(lowercase_ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(lowercase_ ): _lowerCamelCase , _lowerCamelCase = line.rstrip().split('''\t''' ) _lowerCamelCase = index return entity_vocab if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) __SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
623
1
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''BAAI/AltCLIP''': '''https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json''', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str = 'altclip_text_model' def __init__( self , lowerCamelCase__=2_5_0_0_0_2 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=4_0_9_6 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_4 , lowerCamelCase__=1 , lowerCamelCase__=0.0_2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-05 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , lowerCamelCase__="absolute" , lowerCamelCase__=True , lowerCamelCase__=7_6_8 , **lowerCamelCase__ , ): super().__init__(pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = initializer_factor _lowerCamelCase = layer_norm_eps _lowerCamelCase = position_embedding_type _lowerCamelCase = use_cache _lowerCamelCase = project_dim class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = 'altclip_vision_model' def __init__( self , lowerCamelCase__=7_6_8 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3 , lowerCamelCase__=2_2_4 , lowerCamelCase__=3_2 , lowerCamelCase__="quick_gelu" , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1.0 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = hidden_size _lowerCamelCase = intermediate_size _lowerCamelCase = projection_dim _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = num_channels _lowerCamelCase = patch_size _lowerCamelCase = image_size _lowerCamelCase = initializer_range _lowerCamelCase = initializer_factor _lowerCamelCase = attention_dropout _lowerCamelCase = layer_norm_eps _lowerCamelCase = hidden_act @classmethod def snake_case__ ( cls , lowerCamelCase__ , **lowerCamelCase__ ): cls._set_token_in_kwargs(lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get('''model_type''' ) == "altclip": _lowerCamelCase = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Optional[int] = 'altclip' lowercase__ : str = True def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=7_6_8 , lowerCamelCase__=2.6_5_9_2 , **lowerCamelCase__ ): # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). _lowerCamelCase = kwargs.pop('''text_config_dict''' , lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''vision_config_dict''' , lowerCamelCase__ ) super().__init__(**lowerCamelCase__ ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: _lowerCamelCase = {} # This is the complete result when using `text_config_dict`. _lowerCamelCase = AltCLIPTextConfig(**lowerCamelCase__ ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: _lowerCamelCase = ( F"""`{key}` is found in both `text_config_dict` and `text_config` but with different values. """ F"""The value `text_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: _lowerCamelCase = ( F"""`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The """ F"""value `text_config[\"{key}\"]` will be overriden.""" ) logger.warning(lowerCamelCase__ ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: _lowerCamelCase = {} # This is the complete result when using `vision_config_dict`. _lowerCamelCase = AltCLIPVisionConfig(**lowerCamelCase__ ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: _lowerCamelCase = { str(lowerCamelCase__ ): value for key, value in _vision_config_dict['''id2label'''].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: _lowerCamelCase = ( F"""`{key}` is found in both `vision_config_dict` and `vision_config` but with different """ F"""values. The value `vision_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: _lowerCamelCase = ( F"""`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. """ F"""The value `vision_config[\"{key}\"]` will be overriden.""" ) logger.warning(lowerCamelCase__ ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: _lowerCamelCase = {} logger.info('''`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values.''' ) if vision_config is None: _lowerCamelCase = {} logger.info('''`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values.''' ) _lowerCamelCase = AltCLIPTextConfig(**lowerCamelCase__ ) _lowerCamelCase = AltCLIPVisionConfig(**lowerCamelCase__ ) _lowerCamelCase = projection_dim _lowerCamelCase = logit_scale_init_value _lowerCamelCase = 1.0 @classmethod def snake_case__ ( cls , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.text_config.to_dict() _lowerCamelCase = self.vision_config.to_dict() _lowerCamelCase = self.__class__.model_type return output
623
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=3_0 , lowerCamelCase__=2 , lowerCamelCase__=3 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=0.6 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = mask_ratio _lowerCamelCase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) _lowerCamelCase = (image_size // patch_size) ** 2 _lowerCamelCase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEModel(config=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) # expected sequence length = num_patches _lowerCamelCase = (self.image_size // self.patch_size) ** 2 _lowerCamelCase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images _lowerCamelCase = 1 _lowerCamelCase = TFViTMAEForPreTraining(lowerCamelCase__ ) _lowerCamelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowerCamelCase = model(lowerCamelCase__ , training=lowerCamelCase__ ) _lowerCamelCase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () lowercase__ : Dict = {'feature-extraction': TFViTMAEModel} if is_tf_available() else {} lowercase__ : Optional[Any] = False lowercase__ : Union[str, Any] = False lowercase__ : str = False lowercase__ : List[str] = False def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , tf.keras.layers.Layer ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = copy.deepcopy(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = outputs_dict[0].numpy() _lowerCamelCase = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def snake_case__ ( self ): # make the mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase__ ): _lowerCamelCase = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase__ ): _lowerCamelCase = v.numpy() else: _lowerCamelCase = np.array(lowerCamelCase__ ) return inputs_np_dict for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = prepare_numpy_arrays(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # make masks reproducible np.random.seed(2 ) _lowerCamelCase = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument _lowerCamelCase = tf_noise super().check_pt_tf_models(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase__ ) if module_member_name.endswith('''MainLayer''' ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len('''MainLayer''' )] == model_class.__name__[: -len('''Model''' )] for module_member in (getattr(lowerCamelCase__ , lowerCamelCase__ ),) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase__ , '''_keras_serializable''' , lowerCamelCase__ ) } _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) _lowerCamelCase = tf.convert_to_tensor(lowerCamelCase__ ) inputs_dict.update({'''noise''': noise} ) for main_layer_class in tf_main_layer_classes: _lowerCamelCase = main_layer_class(lowerCamelCase__ ) _lowerCamelCase = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } _lowerCamelCase = tf.keras.Model(lowerCamelCase__ , outputs=main_layer(lowerCamelCase__ ) ) _lowerCamelCase = model(lowerCamelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''keras_model.h5''' ) model.save(lowerCamelCase__ ) _lowerCamelCase = tf.keras.models.load_model( lowerCamelCase__ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase__ , tf.keras.Model ) _lowerCamelCase = model(lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = outputs.last_hidden_state.numpy() _lowerCamelCase = 0 else: _lowerCamelCase = outputs.logits.numpy() _lowerCamelCase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = model_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) if model_class.__name__ == "TFViTMAEModel": _lowerCamelCase = after_outputs['''last_hidden_state'''].numpy() _lowerCamelCase = 0 else: _lowerCamelCase = after_outputs['''logits'''].numpy() _lowerCamelCase = 0 _lowerCamelCase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase__ , 1e-5 ) def snake_case__ ( self ): # make mask reproducible np.random.seed(2 ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = int((config.image_size // config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , noise=lowerCamelCase__ ) _lowerCamelCase = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase__ ) _lowerCamelCase = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config _lowerCamelCase = model_class.from_config(model.config ) _lowerCamelCase = new_model(lowerCamelCase__ ) # Build model new_model.set_weights(model.get_weights() ) _lowerCamelCase = new_model(lowerCamelCase__ , noise=lowerCamelCase__ ) self.assert_outputs_same(lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def snake_case__ ( self ): pass @slow def snake_case__ ( self ): _lowerCamelCase = TFViTMAEModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def snake_case__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) _lowerCamelCase = TFViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''tf''' ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) _lowerCamelCase = ViTMAEConfig() _lowerCamelCase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) _lowerCamelCase = np.random.uniform(size=(1, num_patches) ) # forward pass _lowerCamelCase = model(**lowerCamelCase__ , noise=lowerCamelCase__ ) # verify the logits _lowerCamelCase = tf.convert_to_tensor([1, 1_9_6, 7_6_8] ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = tf.convert_to_tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
1
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): if len(lowerCamelCase__ ) != degree + 1: raise ValueError( '''The number of coefficients should be equal to the degree + 1.''' ) _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = degree def __add__( self , lowerCamelCase__ ): if self.degree > polynomial_a.degree: _lowerCamelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , lowerCamelCase__ ) else: _lowerCamelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , lowerCamelCase__ ) def __sub__( self , lowerCamelCase__ ): return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self ): return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self , lowerCamelCase__ ): _lowerCamelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): _lowerCamelCase = '''''' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self ): return self.__str__() def snake_case__ ( self ): _lowerCamelCase = [0] * self.degree for i in range(self.degree ): _lowerCamelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ = 0 ): _lowerCamelCase = [0] * (self.degree + 2) _lowerCamelCase = constant for i in range(self.degree + 1 ): _lowerCamelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , lowerCamelCase__ ) def __eq__( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self , lowerCamelCase__ ): return not self.__eq__(lowerCamelCase__ )
623
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
1
"""simple docstring""" import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency __SCREAMING_SNAKE_CASE : List[Any] = { '''E''': 12.70, '''T''': 9.06, '''A''': 8.17, '''O''': 7.51, '''I''': 6.97, '''N''': 6.75, '''S''': 6.33, '''H''': 6.09, '''R''': 5.99, '''D''': 4.25, '''L''': 4.03, '''C''': 2.78, '''U''': 2.76, '''M''': 2.41, '''W''': 2.36, '''F''': 2.23, '''G''': 2.02, '''Y''': 1.97, '''P''': 1.93, '''B''': 1.29, '''V''': 0.98, '''K''': 0.77, '''J''': 0.15, '''X''': 0.15, '''Q''': 0.10, '''Z''': 0.07, } __SCREAMING_SNAKE_CASE : str = '''ETAOINSHRDLCUMWFGYPBVKJXQZ''' __SCREAMING_SNAKE_CASE : Dict = '''ABCDEFGHIJKLMNOPQRSTUVWXYZ''' def lowerCAmelCase_( lowercase_ : str ) -> dict[str, int]: _lowerCamelCase = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def lowerCAmelCase_( lowercase_ : tuple ) -> str: return x[0] def lowerCAmelCase_( lowercase_ : str ) -> str: _lowerCamelCase = get_letter_count(lowercase_ ) _lowerCamelCase = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(lowercase_ ) _lowerCamelCase = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=lowercase_ ) _lowerCamelCase = ''''''.join(freq_to_letter[freq] ) _lowerCamelCase = list(freq_to_letter_str.items() ) freq_pairs.sort(key=lowercase_ , reverse=lowercase_ ) _lowerCamelCase = [freq_pair[1] for freq_pair in freq_pairs] return "".join(lowercase_ ) def lowerCAmelCase_( lowercase_ : str ) -> int: _lowerCamelCase = get_frequency_order(lowercase_ ) _lowerCamelCase = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
623
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=3.6 ): _lowerCamelCase = tokenizer _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = dataset _lowerCamelCase = seq_length _lowerCamelCase = seq_length * chars_per_token * num_of_sequences def __iter__( self ): _lowerCamelCase = iter(self.dataset ) _lowerCamelCase = True while more_examples: _lowerCamelCase , _lowerCamelCase = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(lowerCamelCase__ )['''content'''] ) buffer_len += len(buffer[-1] ) except StopIteration: _lowerCamelCase = False break _lowerCamelCase = tokenizer(lowerCamelCase__ , truncation=lowerCamelCase__ )['''input_ids'''] _lowerCamelCase = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(lowerCamelCase__ ) , self.seq_length ): _lowerCamelCase = all_token_ids[i : i + self.seq_length] if len(lowerCamelCase__ ) == self.seq_length: yield torch.tensor(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : Any ) -> Optional[Any]: _lowerCamelCase = {'''streaming''': True} _lowerCamelCase = load_dataset(args.dataset_name , split='''train''' , **lowercase_ ) _lowerCamelCase = ConstantLengthDataset(lowercase_ , lowercase_ , seq_length=args.seq_length ) _lowerCamelCase = DataLoader(lowercase_ , batch_size=args.batch_size ) return eval_dataloader def lowerCAmelCase_( lowercase_ : Tuple ) -> str: model.eval() _lowerCamelCase = [] for step, batch in enumerate(lowercase_ ): with torch.no_grad(): _lowerCamelCase = model(lowercase_ , labels=lowercase_ ) _lowerCamelCase = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(lowercase_ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break _lowerCamelCase = torch.mean(torch.cat(lowercase_ ) ) try: _lowerCamelCase = torch.exp(lowercase_ ) except OverflowError: _lowerCamelCase = float('''inf''' ) return loss.item(), perplexity.item() # Setup Accelerator __SCREAMING_SNAKE_CASE : Dict = Accelerator() # Parse configuration __SCREAMING_SNAKE_CASE : Tuple = HfArgumentParser(EvaluationArguments) __SCREAMING_SNAKE_CASE : Optional[int] = parser.parse_args() set_seed(args.seed) # Logging __SCREAMING_SNAKE_CASE : Optional[Any] = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer __SCREAMING_SNAKE_CASE : str = AutoModelForCausalLM.from_pretrained(args.model_ckpt) __SCREAMING_SNAKE_CASE : Union[str, Any] = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader __SCREAMING_SNAKE_CASE : str = create_dataloader(args) # Prepare everything with our `accelerator`. __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = evaluate(args) logger.info(F"""loss/eval: {eval_loss}, perplexity: {perplexity}""")
623
1
"""simple docstring""" from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = 1_3 _lowerCamelCase = 7 _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = 9_9 _lowerCamelCase = 3_8_4 _lowerCamelCase = 2 _lowerCamelCase = 4 _lowerCamelCase = 3_7 _lowerCamelCase = '''gelu''' _lowerCamelCase = 0.1 _lowerCamelCase = 0.1 _lowerCamelCase = 5_1_2 _lowerCamelCase = 1_6 _lowerCamelCase = 2 _lowerCamelCase = 0.0_2 _lowerCamelCase = 3 _lowerCamelCase = 4 _lowerCamelCase = 1_2_8 _lowerCamelCase = 2 _lowerCamelCase = 9 _lowerCamelCase = 1 _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=lowerCamelCase__ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFConvBertModel(config=lowerCamelCase__ ) _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _lowerCamelCase = [input_ids, input_mask] _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFConvBertForMaskedLM(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = TFConvBertForSequenceClassification(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_choices _lowerCamelCase = TFConvBertForMultipleChoice(config=lowerCamelCase__ ) _lowerCamelCase = tf.tile(tf.expand_dims(lowerCamelCase__ , 1 ) , (1, self.num_choices, 1) ) _lowerCamelCase = tf.tile(tf.expand_dims(lowerCamelCase__ , 1 ) , (1, self.num_choices, 1) ) _lowerCamelCase = tf.tile(tf.expand_dims(lowerCamelCase__ , 1 ) , (1, self.num_choices, 1) ) _lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = TFConvBertForTokenClassification(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFConvBertForQuestionAnswering(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) lowercase__ : List[str] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) lowercase__ : List[str] = False lowercase__ : int = False lowercase__ : int = False def snake_case__ ( self ): _lowerCamelCase = TFConvBertModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = True _lowerCamelCase = True if hasattr(lowerCamelCase__ , '''use_cache''' ): _lowerCamelCase = True _lowerCamelCase = getattr(self.model_tester , '''encoder_seq_length''' , self.model_tester.seq_length ) _lowerCamelCase = getattr(self.model_tester , '''key_length''' , lowerCamelCase__ ) for model_class in self.all_model_classes: _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = len(model(lowerCamelCase__ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ , saved_model=lowerCamelCase__ ) _lowerCamelCase = os.path.join(lowerCamelCase__ , '''saved_model''' , '''1''' ) _lowerCamelCase = tf.keras.models.load_model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) if self.is_encoder_decoder: _lowerCamelCase = outputs['''encoder_hidden_states'''] _lowerCamelCase = outputs['''encoder_attentions'''] else: _lowerCamelCase = outputs['''hidden_states'''] _lowerCamelCase = outputs['''attentions'''] self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) _lowerCamelCase = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def snake_case__ ( self ): _lowerCamelCase = TFConvBertModel.from_pretrained('''YituTech/conv-bert-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = True _lowerCamelCase = getattr(self.model_tester , '''decoder_seq_length''' , self.model_tester.seq_length ) _lowerCamelCase = getattr(self.model_tester , '''encoder_seq_length''' , self.model_tester.seq_length ) _lowerCamelCase = getattr(self.model_tester , '''key_length''' , lowerCamelCase__ ) _lowerCamelCase = getattr(self.model_tester , '''key_length''' , lowerCamelCase__ ) def check_decoder_attentions_output(lowerCamelCase__ ): _lowerCamelCase = len(lowerCamelCase__ ) self.assertEqual(out_len % 2 , 0 ) _lowerCamelCase = outputs.decoder_attentions self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(lowerCamelCase__ ): _lowerCamelCase = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(lowerCamelCase__ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: _lowerCamelCase = True _lowerCamelCase = False _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = model(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = len(lowerCamelCase__ ) self.assertEqual(config.output_hidden_states , lowerCamelCase__ ) check_encoder_attentions_output(lowerCamelCase__ ) if self.is_encoder_decoder: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = model(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(config.output_hidden_states , lowerCamelCase__ ) check_decoder_attentions_output(lowerCamelCase__ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _lowerCamelCase = True _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = model(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(config.output_hidden_states , lowerCamelCase__ ) check_encoder_attentions_output(lowerCamelCase__ ) # Check attention is always last and order is fine _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = model(self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(lowerCamelCase__ ) ) self.assertEqual(model.config.output_hidden_states , lowerCamelCase__ ) check_encoder_attentions_output(lowerCamelCase__ ) @require_tf class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = TFConvBertModel.from_pretrained('''YituTech/conv-bert-base''' ) _lowerCamelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) _lowerCamelCase = model(lowerCamelCase__ )[0] _lowerCamelCase = [1, 6, 7_6_8] self.assertEqual(output.shape , lowerCamelCase__ ) _lowerCamelCase = tf.constant( [ [ [-0.0_3_4_7_5_4_9_3, -0.4_6_8_6_0_3_4, -0.3_0_6_3_8_8_3_2], [0.2_2_6_3_7_2_4_8, -0.2_6_9_8_8_6_4_6, -0.7_4_2_3_4_2_4], [0.1_0_3_2_4_8_6_8, -0.4_5_0_1_3_5_0_8, -0.5_8_2_8_0_7_8_4], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowerCamelCase__ , atol=1e-4 )
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : np.ndarray , lowercase_ : float = 1e-12 , lowercase_ : int = 1_00 , ) -> tuple[float, np.ndarray]: assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[1] # Ensure proper dimensionality. assert np.shape(lowercase_ )[0] == np.shape(lowercase_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(lowercase_ ) == np.iscomplexobj(lowercase_ ) _lowerCamelCase = np.iscomplexobj(lowercase_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(lowercase_ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _lowerCamelCase = False _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 1e12 while not convergence: # Multiple matrix by the vector. _lowerCamelCase = np.dot(lowercase_ , lowercase_ ) # Normalize the resulting output vector. _lowerCamelCase = w / np.linalg.norm(lowercase_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _lowerCamelCase = vector.conj().T if is_complex else vector.T _lowerCamelCase = np.dot(lowercase_ , np.dot(lowercase_ , lowercase_ ) ) # Check convergence. _lowerCamelCase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _lowerCamelCase = True _lowerCamelCase = lambda_ if is_complex: _lowerCamelCase = np.real(lambda_ ) return lambda_, vector def lowerCAmelCase_( ) -> None: _lowerCamelCase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) _lowerCamelCase = np.array([41, 4, 20] ) _lowerCamelCase = real_input_matrix.astype(np.complexaaa ) _lowerCamelCase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _lowerCamelCase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": _lowerCamelCase = real_input_matrix _lowerCamelCase = real_vector elif problem_type == "complex": _lowerCamelCase = complex_input_matrix _lowerCamelCase = complex_vector # Our implementation. _lowerCamelCase , _lowerCamelCase = power_iteration(lowercase_ , lowercase_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _lowerCamelCase , _lowerCamelCase = np.linalg.eigh(lowercase_ ) # Last eigenvalue is the maximum one. _lowerCamelCase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _lowerCamelCase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(lowercase_ ) - np.abs(lowercase_ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
623
1
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = (PNDMScheduler,) lowercase__ : int = (('num_inference_steps', 50),) def snake_case__ ( self , **lowerCamelCase__ ): _lowerCamelCase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0_0_0_1, '''beta_end''': 0.0_2, '''beta_schedule''': '''linear''', } config.update(**lowerCamelCase__ ) return config def snake_case__ ( self , lowerCamelCase__=0 , **lowerCamelCase__ ): _lowerCamelCase = dict(self.forward_default_kwargs ) _lowerCamelCase = kwargs.pop('''num_inference_steps''' , lowerCamelCase__ ) _lowerCamelCase = self.dummy_sample _lowerCamelCase = 0.1 * sample _lowerCamelCase = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _lowerCamelCase = self.get_scheduler_config(**lowerCamelCase__ ) _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals _lowerCamelCase = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) _lowerCamelCase = scheduler_class.from_pretrained(lowerCamelCase__ ) new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals _lowerCamelCase = dummy_past_residuals[:] _lowerCamelCase = scheduler.step_prk(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _lowerCamelCase = new_scheduler.step_prk(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" _lowerCamelCase = scheduler.step_plms(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _lowerCamelCase = new_scheduler.step_plms(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def snake_case__ ( self ): pass def snake_case__ ( self , lowerCamelCase__=0 , **lowerCamelCase__ ): _lowerCamelCase = dict(self.forward_default_kwargs ) _lowerCamelCase = kwargs.pop('''num_inference_steps''' , lowerCamelCase__ ) _lowerCamelCase = self.dummy_sample _lowerCamelCase = 0.1 * sample _lowerCamelCase = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _lowerCamelCase = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCamelCase__ ) _lowerCamelCase = scheduler_class.from_pretrained(lowerCamelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCamelCase__ ) # copy over dummy past residual (must be after setting timesteps) _lowerCamelCase = dummy_past_residuals[:] _lowerCamelCase = scheduler.step_prk(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _lowerCamelCase = new_scheduler.step_prk(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" _lowerCamelCase = scheduler.step_plms(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _lowerCamelCase = new_scheduler.step_plms(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def snake_case__ ( self , **lowerCamelCase__ ): _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config(**lowerCamelCase__ ) _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) _lowerCamelCase = 1_0 _lowerCamelCase = self.dummy_model() _lowerCamelCase = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase__ ) for i, t in enumerate(scheduler.prk_timesteps ): _lowerCamelCase = model(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = scheduler.step_prk(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): _lowerCamelCase = model(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = scheduler.step_plms(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample return sample def snake_case__ ( self ): _lowerCamelCase = dict(self.forward_default_kwargs ) _lowerCamelCase = kwargs.pop('''num_inference_steps''' , lowerCamelCase__ ) for scheduler_class in self.scheduler_classes: _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) _lowerCamelCase = self.dummy_sample _lowerCamelCase = 0.1 * sample if num_inference_steps is not None and hasattr(lowerCamelCase__ , '''set_timesteps''' ): scheduler.set_timesteps(lowerCamelCase__ ) elif num_inference_steps is not None and not hasattr(lowerCamelCase__ , '''set_timesteps''' ): _lowerCamelCase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _lowerCamelCase = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] _lowerCamelCase = dummy_past_residuals[:] _lowerCamelCase = scheduler.step_prk(lowerCamelCase__ , 0 , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _lowerCamelCase = scheduler.step_prk(lowerCamelCase__ , 1 , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _lowerCamelCase = scheduler.step_plms(lowerCamelCase__ , 0 , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample _lowerCamelCase = scheduler.step_plms(lowerCamelCase__ , 1 , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def snake_case__ ( self ): for timesteps in [1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=lowerCamelCase__ ) def snake_case__ ( self ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowerCamelCase__ ) _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config(steps_offset=1 ) _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(1_0 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [9_0_1, 8_5_1, 8_5_1, 8_0_1, 8_0_1, 7_5_1, 7_5_1, 7_0_1, 7_0_1, 6_5_1, 6_5_1, 6_0_1, 6_0_1, 5_0_1, 4_0_1, 3_0_1, 2_0_1, 1_0_1, 1] ) , ) def snake_case__ ( self ): for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1] , [0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=lowerCamelCase__ , beta_end=lowerCamelCase__ ) def snake_case__ ( self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowerCamelCase__ ) def snake_case__ ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase__ ) def snake_case__ ( self ): for t in [1, 5, 1_0]: self.check_over_forward(time_step=lowerCamelCase__ ) def snake_case__ ( self ): for t, num_inference_steps in zip([1, 5, 1_0] , [1_0, 5_0, 1_0_0] ): self.check_over_forward(num_inference_steps=lowerCamelCase__ ) def snake_case__ ( self ): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 _lowerCamelCase = 2_7 for scheduler_class in self.scheduler_classes: _lowerCamelCase = self.dummy_sample _lowerCamelCase = 0.1 * sample _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) scheduler.set_timesteps(lowerCamelCase__ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): _lowerCamelCase = scheduler.step_prk(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample def snake_case__ ( self ): with self.assertRaises(lowerCamelCase__ ): _lowerCamelCase = self.scheduler_classes[0] _lowerCamelCase = self.get_scheduler_config() _lowerCamelCase = scheduler_class(**lowerCamelCase__ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def snake_case__ ( self ): _lowerCamelCase = self.full_loop() _lowerCamelCase = torch.sum(torch.abs(lowerCamelCase__ ) ) _lowerCamelCase = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 1_9_8.1_3_1_8 ) < 1e-2 assert abs(result_mean.item() - 0.2_5_8_0 ) < 1e-3 def snake_case__ ( self ): _lowerCamelCase = self.full_loop(prediction_type='''v_prediction''' ) _lowerCamelCase = torch.sum(torch.abs(lowerCamelCase__ ) ) _lowerCamelCase = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 6_7.3_9_8_6 ) < 1e-2 assert abs(result_mean.item() - 0.0_8_7_8 ) < 1e-3 def snake_case__ ( self ): # We specify different beta, so that the first alpha is 0.99 _lowerCamelCase = self.full_loop(set_alpha_to_one=lowerCamelCase__ , beta_start=0.0_1 ) _lowerCamelCase = torch.sum(torch.abs(lowerCamelCase__ ) ) _lowerCamelCase = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 2_3_0.0_3_9_9 ) < 1e-2 assert abs(result_mean.item() - 0.2_9_9_5 ) < 1e-3 def snake_case__ ( self ): # We specify different beta, so that the first alpha is 0.99 _lowerCamelCase = self.full_loop(set_alpha_to_one=lowerCamelCase__ , beta_start=0.0_1 ) _lowerCamelCase = torch.sum(torch.abs(lowerCamelCase__ ) ) _lowerCamelCase = torch.mean(torch.abs(lowerCamelCase__ ) ) assert abs(result_sum.item() - 1_8_6.9_4_8_2 ) < 1e-2 assert abs(result_mean.item() - 0.2_4_3_4 ) < 1e-3
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_speecht5''': [ '''SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP''', '''SpeechT5Config''', '''SpeechT5HifiGanConfig''', ], '''feature_extraction_speecht5''': ['''SpeechT5FeatureExtractor'''], '''processing_speecht5''': ['''SpeechT5Processor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ['''SpeechT5Tokenizer'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Any = [ '''SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SpeechT5ForSpeechToText''', '''SpeechT5ForSpeechToSpeech''', '''SpeechT5ForTextToSpeech''', '''SpeechT5Model''', '''SpeechT5PreTrainedModel''', '''SpeechT5HifiGan''', ] if TYPE_CHECKING: from .configuration_speechta import ( SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP, SPEECHT5_PRETRAINED_HIFIGAN_CONFIG_ARCHIVE_MAP, SpeechTaConfig, SpeechTaHifiGanConfig, ) from .feature_extraction_speechta import SpeechTaFeatureExtractor from .processing_speechta import SpeechTaProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speechta import SpeechTaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speechta import ( SPEECHT5_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaModel, SpeechTaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableDiffusionUpscalePipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def snake_case__ ( self ): _lowerCamelCase = 1 _lowerCamelCase = 3 _lowerCamelCase = (3_2, 3_2) _lowerCamelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCamelCase__ ) return image @property def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=7 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , attention_head_dim=8 , use_linear_projection=lowerCamelCase__ , only_cross_attention=(True, True, False) , num_class_embeds=1_0_0 , ) return model @property def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) return model @property def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=5_1_2 , ) return CLIPTextModel(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.dummy_cond_unet_upscale _lowerCamelCase = DDPMScheduler() _lowerCamelCase = DDIMScheduler(prediction_type='''v_prediction''' ) _lowerCamelCase = self.dummy_vae _lowerCamelCase = self.dummy_text_encoder _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) _lowerCamelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCamelCase = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('''RGB''' ).resize((6_4, 6_4) ) # make sure here that pndm scheduler skips prk _lowerCamelCase = StableDiffusionUpscalePipeline( unet=lowerCamelCase__ , low_res_scheduler=lowerCamelCase__ , scheduler=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , max_noise_level=3_5_0 , ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger''' _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) _lowerCamelCase = sd_pipe( [prompt] , image=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type='''np''' , ) _lowerCamelCase = output.images _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) _lowerCamelCase = sd_pipe( [prompt] , image=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type='''np''' , return_dict=lowerCamelCase__ , )[0] _lowerCamelCase = image[0, -3:, -3:, -1] _lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] _lowerCamelCase = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) _lowerCamelCase = np.array([0.3_1_1_3, 0.3_9_1_0, 0.4_2_7_2, 0.4_8_5_9, 0.5_0_6_1, 0.4_6_5_2, 0.5_3_6_2, 0.5_7_1_5, 0.5_6_6_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.dummy_cond_unet_upscale _lowerCamelCase = DDPMScheduler() _lowerCamelCase = DDIMScheduler(prediction_type='''v_prediction''' ) _lowerCamelCase = self.dummy_vae _lowerCamelCase = self.dummy_text_encoder _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) _lowerCamelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCamelCase = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('''RGB''' ).resize((6_4, 6_4) ) # make sure here that pndm scheduler skips prk _lowerCamelCase = StableDiffusionUpscalePipeline( unet=lowerCamelCase__ , low_res_scheduler=lowerCamelCase__ , scheduler=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , max_noise_level=3_5_0 , ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger''' _lowerCamelCase = sd_pipe( 2 * [prompt] , image=2 * [low_res_image] , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type='''np''' , ) _lowerCamelCase = output.images assert image.shape[0] == 2 _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) _lowerCamelCase = sd_pipe( [prompt] , image=lowerCamelCase__ , generator=lowerCamelCase__ , num_images_per_prompt=2 , guidance_scale=6.0 , noise_level=2_0 , num_inference_steps=2 , output_type='''np''' , ) _lowerCamelCase = output.images assert image.shape[0] == 2 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def snake_case__ ( self ): _lowerCamelCase = self.dummy_cond_unet_upscale _lowerCamelCase = DDPMScheduler() _lowerCamelCase = DDIMScheduler(prediction_type='''v_prediction''' ) _lowerCamelCase = self.dummy_vae _lowerCamelCase = self.dummy_text_encoder _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) _lowerCamelCase = self.dummy_image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCamelCase = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('''RGB''' ).resize((6_4, 6_4) ) # put models in fp16, except vae as it overflows in fp16 _lowerCamelCase = unet.half() _lowerCamelCase = text_encoder.half() # make sure here that pndm scheduler skips prk _lowerCamelCase = StableDiffusionUpscalePipeline( unet=lowerCamelCase__ , low_res_scheduler=lowerCamelCase__ , scheduler=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , max_noise_level=3_5_0 , ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = '''A painting of a squirrel eating a burger''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = sd_pipe( [prompt] , image=lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=2 , output_type='''np''' , ).images _lowerCamelCase = low_res_image.size[0] * 4 assert image.shape == (1, expected_height_width, expected_height_width, 3) @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self ): _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) _lowerCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale''' '''/upsampled_cat.npy''' ) _lowerCamelCase = '''stabilityai/stable-diffusion-x4-upscaler''' _lowerCamelCase = StableDiffusionUpscalePipeline.from_pretrained(lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() _lowerCamelCase = '''a cat sitting on a park bench''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , image=lowerCamelCase__ , generator=lowerCamelCase__ , output_type='''np''' , ) _lowerCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 1e-3 def snake_case__ ( self ): _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) _lowerCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-upscale''' '''/upsampled_cat_fp16.npy''' ) _lowerCamelCase = '''stabilityai/stable-diffusion-x4-upscaler''' _lowerCamelCase = StableDiffusionUpscalePipeline.from_pretrained( lowerCamelCase__ , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() _lowerCamelCase = '''a cat sitting on a park bench''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , image=lowerCamelCase__ , generator=lowerCamelCase__ , output_type='''np''' , ) _lowerCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image ).max() < 5e-1 def snake_case__ ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-upscale/low_res_cat.png''' ) _lowerCamelCase = '''stabilityai/stable-diffusion-x4-upscaler''' _lowerCamelCase = StableDiffusionUpscalePipeline.from_pretrained( lowerCamelCase__ , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _lowerCamelCase = '''a cat sitting on a park bench''' _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( prompt=lowerCamelCase__ , image=lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=5 , output_type='''np''' , ) _lowerCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.9 GB is allocated assert mem_bytes < 2.9 * 1_0**9
623
"""simple docstring""" from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __SCREAMING_SNAKE_CASE : List[str] = numpy.array([0, 0]) __SCREAMING_SNAKE_CASE : Optional[Any] = numpy.array([0.5, 0.866_0254]) __SCREAMING_SNAKE_CASE : Tuple = numpy.array([1, 0]) __SCREAMING_SNAKE_CASE : List[Any] = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] , lowercase_ : int ) -> list[numpy.ndarray]: _lowerCamelCase = initial_vectors for _ in range(lowercase_ ): _lowerCamelCase = iteration_step(lowercase_ ) return vectors def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> list[numpy.ndarray]: _lowerCamelCase = [] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase = vectors[i + 1] new_vectors.append(lowercase_ ) _lowerCamelCase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def lowerCAmelCase_( lowercase_ : numpy.ndarray , lowercase_ : float ) -> numpy.ndarray: _lowerCamelCase = numpy.radians(lowercase_ ) _lowerCamelCase , _lowerCamelCase = numpy.cos(lowercase_ ), numpy.sin(lowercase_ ) _lowerCamelCase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : list[numpy.ndarray] ) -> None: _lowerCamelCase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase , _lowerCamelCase = zip(*lowercase_ ) plt.plot(lowercase_ , lowercase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() __SCREAMING_SNAKE_CASE : str = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
623
1
"""simple docstring""" import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('''At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training''') # TF training parameters __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : Dict = False def lowerCAmelCase_( lowercase_ : Namespace ) -> Optional[int]: return TrainCommand(lowercase_ ) class lowerCamelCase_( A__ ): '''simple docstring''' @staticmethod def snake_case__ ( lowerCamelCase__ ): _lowerCamelCase = parser.add_parser('''train''' , help='''CLI tool to train a model on a task.''' ) train_parser.add_argument( '''--train_data''' , type=lowerCamelCase__ , required=lowerCamelCase__ , help='''path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.''' , ) train_parser.add_argument( '''--column_label''' , type=lowerCamelCase__ , default=0 , help='''Column of the dataset csv file with example labels.''' ) train_parser.add_argument( '''--column_text''' , type=lowerCamelCase__ , default=1 , help='''Column of the dataset csv file with example texts.''' ) train_parser.add_argument( '''--column_id''' , type=lowerCamelCase__ , default=2 , help='''Column of the dataset csv file with example ids.''' ) train_parser.add_argument( '''--skip_first_row''' , action='''store_true''' , help='''Skip the first row of the csv file (headers).''' ) train_parser.add_argument('''--validation_data''' , type=lowerCamelCase__ , default='''''' , help='''path to validation dataset.''' ) train_parser.add_argument( '''--validation_split''' , type=lowerCamelCase__ , default=0.1 , help='''if validation dataset is not provided, fraction of train dataset to use as validation dataset.''' , ) train_parser.add_argument('''--output''' , type=lowerCamelCase__ , default='''./''' , help='''path to saved the trained model.''' ) train_parser.add_argument( '''--task''' , type=lowerCamelCase__ , default='''text_classification''' , help='''Task to train the model on.''' ) train_parser.add_argument( '''--model''' , type=lowerCamelCase__ , default='''bert-base-uncased''' , help='''Model\'s name or path to stored model.''' ) train_parser.add_argument('''--train_batch_size''' , type=lowerCamelCase__ , default=3_2 , help='''Batch size for training.''' ) train_parser.add_argument('''--valid_batch_size''' , type=lowerCamelCase__ , default=6_4 , help='''Batch size for validation.''' ) train_parser.add_argument('''--learning_rate''' , type=lowerCamelCase__ , default=3e-5 , help='''Learning rate.''' ) train_parser.add_argument('''--adam_epsilon''' , type=lowerCamelCase__ , default=1e-08 , help='''Epsilon for Adam optimizer.''' ) train_parser.set_defaults(func=lowerCamelCase__ ) def __init__( self , lowerCamelCase__ ): _lowerCamelCase = logging.get_logger('''transformers-cli/training''' ) _lowerCamelCase = '''tf''' if is_tf_available() else '''torch''' os.makedirs(args.output , exist_ok=lowerCamelCase__ ) _lowerCamelCase = args.output _lowerCamelCase = args.column_label _lowerCamelCase = args.column_text _lowerCamelCase = args.column_id self.logger.info(F"""Loading {args.task} pipeline for {args.model}""" ) if args.task == "text_classification": _lowerCamelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(F"""Loading dataset from {args.train_data}""" ) _lowerCamelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) _lowerCamelCase = None if args.validation_data: self.logger.info(F"""Loading validation dataset from {args.validation_data}""" ) _lowerCamelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) _lowerCamelCase = args.validation_split _lowerCamelCase = args.train_batch_size _lowerCamelCase = args.valid_batch_size _lowerCamelCase = args.learning_rate _lowerCamelCase = args.adam_epsilon def snake_case__ ( self ): if self.framework == "tf": return self.run_tf() return self.run_torch() def snake_case__ ( self ): raise NotImplementedError def snake_case__ ( self ): self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
623
"""simple docstring""" from typing import Any class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = self.head while temp is not None: print(temp.data , end=''' ''' ) _lowerCamelCase = temp.next print() def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = Node(lowerCamelCase__ ) _lowerCamelCase = self.head _lowerCamelCase = new_node def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if node_data_a == node_data_a: return else: _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next _lowerCamelCase = self.head while node_a is not None and node_a.data != node_data_a: _lowerCamelCase = node_a.next if node_a is None or node_a is None: return _lowerCamelCase , _lowerCamelCase = node_a.data, node_a.data if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('''After swapping''') ll.print_list()
623
1
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder __SCREAMING_SNAKE_CASE : List[Any] = '''__DUMMY_TRANSFORMERS_USER__''' __SCREAMING_SNAKE_CASE : Union[str, Any] = '''Dummy User''' __SCREAMING_SNAKE_CASE : Tuple = '''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''' __SCREAMING_SNAKE_CASE : List[str] = '''https://hub-ci.huggingface.co''' __SCREAMING_SNAKE_CASE : Union[str, Any] = CI_HUB_ENDPOINT + '''/datasets/{repo_id}/resolve/{revision}/{path}''' __SCREAMING_SNAKE_CASE : Tuple = CI_HUB_ENDPOINT + '''/{repo_id}/resolve/{revision}/{filename}''' __SCREAMING_SNAKE_CASE : Dict = Path('''~/.huggingface/hub_ci_token''').expanduser() @pytest.fixture def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> List[Any]: monkeypatch.setattr( '''huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE''' , lowercase_ ) @pytest.fixture def lowerCAmelCase_( lowercase_ : Any ) -> Dict: monkeypatch.setattr('''datasets.config.HF_ENDPOINT''' , lowercase_ ) monkeypatch.setattr('''datasets.config.HUB_DATASETS_URL''' , lowercase_ ) @pytest.fixture def lowerCAmelCase_( lowercase_ : Any ) -> List[Any]: monkeypatch.setattr('''huggingface_hub.hf_api.HfFolder.path_token''' , lowercase_ ) @pytest.fixture def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : Union[str, Any] ) -> Optional[Any]: HfFolder.save_token(lowercase_ ) yield HfFolder.delete_token() @pytest.fixture(scope='''session''' ) def lowerCAmelCase_( ) -> Dict: return HfApi(endpoint=lowercase_ ) @pytest.fixture(scope='''session''' ) def lowerCAmelCase_( lowercase_ : HfApi ) -> Union[str, Any]: _lowerCamelCase = HfFolder.get_token() HfFolder.save_token(lowercase_ ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(lowercase_ ) @pytest.fixture def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> Dict: def _cleanup_repo(lowercase_ : Tuple ): hf_api.delete_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' ) return _cleanup_repo @pytest.fixture def lowerCAmelCase_( lowercase_ : List[str] ) -> str: @contextmanager def _temporary_repo(lowercase_ : Dict ): try: yield repo_id finally: cleanup_repo(lowercase_ ) return _temporary_repo @pytest.fixture(scope='''session''' ) def lowerCAmelCase_( lowercase_ : HfApi , lowercase_ : str , lowercase_ : Union[str, Any] ) -> Optional[int]: _lowerCamelCase = F"""repo_txt_data-{int(time.time() * 10e3 )}""" _lowerCamelCase = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' , private=lowercase_ ) hf_api.upload_file( token=lowercase_ , path_or_fileobj=str(lowercase_ ) , path_in_repo='''data/text_data.txt''' , repo_id=lowercase_ , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : Union[str, Any] ) -> Tuple: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope='''session''' ) def lowerCAmelCase_( lowercase_ : HfApi , lowercase_ : Any , lowercase_ : List[Any] ) -> Tuple: _lowerCamelCase = F"""repo_zipped_txt_data-{int(time.time() * 10e3 )}""" _lowerCamelCase = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' , private=lowercase_ ) hf_api.upload_file( token=lowercase_ , path_or_fileobj=str(lowercase_ ) , path_in_repo='''data.zip''' , repo_id=lowercase_ , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Union[str, Any] ) -> Union[str, Any]: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope='''session''' ) def lowerCAmelCase_( lowercase_ : HfApi , lowercase_ : Optional[int] , lowercase_ : Union[str, Any] ) -> str: _lowerCamelCase = F"""repo_zipped_img_data-{int(time.time() * 10e3 )}""" _lowerCamelCase = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' , private=lowercase_ ) hf_api.upload_file( token=lowercase_ , path_or_fileobj=str(lowercase_ ) , path_in_repo='''data.zip''' , repo_id=lowercase_ , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(lowercase_ , token=lowercase_ , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def lowerCAmelCase_( lowercase_ : str , lowercase_ : Dict , lowercase_ : List[str] ) -> Optional[int]: return hf_private_dataset_repo_zipped_img_data_
623
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. __SCREAMING_SNAKE_CASE : Optional[Any] = abspath(join(dirname(dirname(__file__)), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[Any]: from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[str] ) -> List[str]: from diffusers.utils.testing_utils import pytest_terminal_summary_main _lowerCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(lowercase_ , id=lowercase_ )
623
1
"""simple docstring""" import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __SCREAMING_SNAKE_CASE : Optional[int] = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = 'AutoTokenizer' lowercase__ : Dict = ['tokenizer'] lowercase__ : str = { 'semantic_prompt': 1, 'coarse_prompt': 2, 'fine_prompt': 2, } def __init__( self , lowerCamelCase__ , lowerCamelCase__=None ): super().__init__(lowerCamelCase__ ) _lowerCamelCase = speaker_embeddings @classmethod def snake_case__ ( cls , lowerCamelCase__ , lowerCamelCase__="speaker_embeddings_path.json" , **lowerCamelCase__ ): if speaker_embeddings_dict_path is not None: _lowerCamelCase = get_file_from_repo( lowerCamelCase__ , lowerCamelCase__ , subfolder=kwargs.pop('''subfolder''' , lowerCamelCase__ ) , cache_dir=kwargs.pop('''cache_dir''' , lowerCamelCase__ ) , force_download=kwargs.pop('''force_download''' , lowerCamelCase__ ) , proxies=kwargs.pop('''proxies''' , lowerCamelCase__ ) , resume_download=kwargs.pop('''resume_download''' , lowerCamelCase__ ) , local_files_only=kwargs.pop('''local_files_only''' , lowerCamelCase__ ) , use_auth_token=kwargs.pop('''use_auth_token''' , lowerCamelCase__ ) , revision=kwargs.pop('''revision''' , lowerCamelCase__ ) , ) if speaker_embeddings_path is None: logger.warning( F"""`{os.path.join(lowerCamelCase__ , lowerCamelCase__ )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) _lowerCamelCase = None else: with open(lowerCamelCase__ ) as speaker_embeddings_json: _lowerCamelCase = json.load(lowerCamelCase__ ) else: _lowerCamelCase = None _lowerCamelCase = AutoTokenizer.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) return cls(tokenizer=lowerCamelCase__ , speaker_embeddings=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="speaker_embeddings_path.json" , lowerCamelCase__="speaker_embeddings" , lowerCamelCase__ = False , **lowerCamelCase__ , ): if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase__ , lowerCamelCase__ , '''v2''' ) , exist_ok=lowerCamelCase__ ) _lowerCamelCase = {} _lowerCamelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _lowerCamelCase = self._load_voice_preset(lowerCamelCase__ ) _lowerCamelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['''repo_or_path'''] , lowerCamelCase__ , F"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=lowerCamelCase__ , ) _lowerCamelCase = os.path.join(lowerCamelCase__ , F"""{prompt_key}_{key}.npy""" ) _lowerCamelCase = tmp_dict with open(os.path.join(lowerCamelCase__ , lowerCamelCase__ ) , '''w''' ) as fp: json.dump(lowerCamelCase__ , lowerCamelCase__ ) super().save_pretrained(lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ = None , **lowerCamelCase__ ): _lowerCamelCase = self.speaker_embeddings[voice_preset] _lowerCamelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) _lowerCamelCase = get_file_from_repo( self.speaker_embeddings.get('''repo_or_path''' , '''/''' ) , voice_preset_paths[key] , subfolder=kwargs.pop('''subfolder''' , lowerCamelCase__ ) , cache_dir=kwargs.pop('''cache_dir''' , lowerCamelCase__ ) , force_download=kwargs.pop('''force_download''' , lowerCamelCase__ ) , proxies=kwargs.pop('''proxies''' , lowerCamelCase__ ) , resume_download=kwargs.pop('''resume_download''' , lowerCamelCase__ ) , local_files_only=kwargs.pop('''local_files_only''' , lowerCamelCase__ ) , use_auth_token=kwargs.pop('''use_auth_token''' , lowerCamelCase__ ) , revision=kwargs.pop('''revision''' , lowerCamelCase__ ) , ) if path is None: raise ValueError( F"""`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) _lowerCamelCase = np.load(lowerCamelCase__ ) return voice_preset_dict def snake_case__ ( self , lowerCamelCase__ = None ): for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__="pt" , lowerCamelCase__=2_5_6 , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=False , **lowerCamelCase__ , ): if voice_preset is not None and not isinstance(lowerCamelCase__ , lowerCamelCase__ ): if ( isinstance(lowerCamelCase__ , lowerCamelCase__ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _lowerCamelCase = self._load_voice_preset(lowerCamelCase__ ) else: if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and not voice_preset.endswith('''.npz''' ): _lowerCamelCase = voice_preset + '''.npz''' _lowerCamelCase = np.load(lowerCamelCase__ ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ ) _lowerCamelCase = self.tokenizer( lowerCamelCase__ , return_tensors=lowerCamelCase__ , padding='''max_length''' , max_length=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , **lowerCamelCase__ , ) if voice_preset is not None: _lowerCamelCase = voice_preset return encoded_text
623
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import KandinskyPipeline, KandinskyPriorPipeline else: from .pipeline_kandinsky import KandinskyPipeline from .pipeline_kandinsky_imgaimg import KandinskyImgaImgPipeline from .pipeline_kandinsky_inpaint import KandinskyInpaintPipeline from .pipeline_kandinsky_prior import KandinskyPriorPipeline, KandinskyPriorPipelineOutput from .text_encoder import MultilingualCLIP
623
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[int] = { '''facebook/s2t-small-librispeech-asr''': ( '''https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json''' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech_to_text } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : int = 'speech_to_text' lowercase__ : List[Any] = ['past_key_values'] lowercase__ : Optional[int] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , lowerCamelCase__=1_0_0_0_0 , lowerCamelCase__=1_2 , lowerCamelCase__=2_0_4_8 , lowerCamelCase__=4 , lowerCamelCase__=6 , lowerCamelCase__=2_0_4_8 , lowerCamelCase__=4 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=2_5_6 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=2 , lowerCamelCase__=True , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , lowerCamelCase__=6_0_0_0 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=2 , lowerCamelCase__=(5, 5) , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=8_0 , lowerCamelCase__=1 , **lowerCamelCase__ , ): _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = encoder_ffn_dim _lowerCamelCase = encoder_layers _lowerCamelCase = encoder_attention_heads _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_attention_heads _lowerCamelCase = dropout _lowerCamelCase = attention_dropout _lowerCamelCase = activation_dropout _lowerCamelCase = activation_function _lowerCamelCase = init_std _lowerCamelCase = encoder_layerdrop _lowerCamelCase = decoder_layerdrop _lowerCamelCase = use_cache _lowerCamelCase = encoder_layers _lowerCamelCase = scale_embedding # scale factor will be sqrt(d_model) if True _lowerCamelCase = max_source_positions _lowerCamelCase = max_target_positions _lowerCamelCase = num_conv_layers _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = conv_channels _lowerCamelCase = input_feat_per_channel _lowerCamelCase = input_channels if len(self.conv_kernel_sizes ) != self.num_conv_layers: raise ValueError( '''Configuration for convolutional module is incorrect. ''' '''It is required that `len(config.conv_kernel_sizes)` == `config.num_conv_layers` ''' F"""but is `len(config.conv_kernel_sizes) = {len(self.conv_kernel_sizes )}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""" ) super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , is_encoder_decoder=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
623
"""simple docstring""" import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=9_9 , lowerCamelCase__=1_3 , lowerCamelCase__=1_6 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=2 , lowerCamelCase__=3_2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=3_0 , lowerCamelCase__=0 , lowerCamelCase__=1 , lowerCamelCase__=2 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = decoder_seq_length # For common tests _lowerCamelCase = self.decoder_seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = d_model _lowerCamelCase = d_model _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_attention_heads _lowerCamelCase = decoder_attention_heads _lowerCamelCase = eos_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = decoder_start_token_id _lowerCamelCase = use_cache _lowerCamelCase = max_position_embeddings _lowerCamelCase = None _lowerCamelCase = decoder_seq_length _lowerCamelCase = 2 _lowerCamelCase = 1 def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) _lowerCamelCase = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): _lowerCamelCase = True _lowerCamelCase = TrOCRDecoder(config=lowerCamelCase__ ).to(lowerCamelCase__ ).eval() _lowerCamelCase = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , use_cache=lowerCamelCase__ ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) ) self.parent.assertTrue(len(lowerCamelCase__ ) == len(lowerCamelCase__ ) + 1 ) _lowerCamelCase = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids _lowerCamelCase = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and _lowerCamelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase = model(lowerCamelCase__ )['''last_hidden_state'''] _lowerCamelCase = model(lowerCamelCase__ , past_key_values=lowerCamelCase__ )['''last_hidden_state'''] # select random slice _lowerCamelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() _lowerCamelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : int = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowercase__ : List[str] = (TrOCRForCausalLM,) if is_torch_available() else () lowercase__ : Tuple = {'text-generation': TrOCRForCausalLM} if is_torch_available() else {} lowercase__ : Dict = True lowercase__ : Optional[Any] = False def snake_case__ ( self ): _lowerCamelCase = TrOCRStandaloneDecoderModelTester(self , is_training=lowerCamelCase__ ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): pass def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*lowerCamelCase__ ) def snake_case__ ( self ): return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self ): pass
623
1
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline __SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): super().__init__() self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) @torch.no_grad() def __call__( self , lowerCamelCase__ = 1 , lowerCamelCase__ = 1_0_0 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = True , ): if audio_length_in_s is None: _lowerCamelCase = self.unet.config.sample_size / self.unet.config.sample_rate _lowerCamelCase = audio_length_in_s * self.unet.config.sample_rate _lowerCamelCase = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( F"""{audio_length_in_s} is too small. Make sure it's bigger or equal to""" F""" {3 * down_scale_factor / self.unet.config.sample_rate}.""" ) _lowerCamelCase = int(lowerCamelCase__ ) if sample_size % down_scale_factor != 0: _lowerCamelCase = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( F"""{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled""" F""" by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising""" ''' process.''' ) _lowerCamelCase = int(lowerCamelCase__ ) _lowerCamelCase = next(iter(self.unet.parameters() ) ).dtype _lowerCamelCase = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(lowerCamelCase__ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _lowerCamelCase = randn_tensor(lowerCamelCase__ , generator=lowerCamelCase__ , device=self.device , dtype=lowerCamelCase__ ) # set step values self.scheduler.set_timesteps(lowerCamelCase__ , device=audio.device ) _lowerCamelCase = self.scheduler.timesteps.to(lowerCamelCase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output _lowerCamelCase = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # 2. compute previous image: x_t -> t_t-1 _lowerCamelCase = self.scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample _lowerCamelCase = audio.clamp(-1 , 1 ).float().cpu().numpy() _lowerCamelCase = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=lowerCamelCase__ )
623
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__=None , **lowerCamelCase__ ): warnings.warn( '''`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` ''' '''instead.''' , lowerCamelCase__ , ) super().__init__(args=lowerCamelCase__ , **lowerCamelCase__ )
623
1
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=1_6 , lowerCamelCase__=3_6 , lowerCamelCase__=6 , lowerCamelCase__=6 , lowerCamelCase__=6 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_input_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = embedding_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_hidden_groups _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self ): return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = AlbertModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = AlbertForPreTraining(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ , sentence_order_label=lowerCamelCase__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = AlbertForMaskedLM(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = AlbertForQuestionAnswering(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , start_positions=lowerCamelCase__ , end_positions=lowerCamelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = AlbertForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = AlbertForTokenClassification(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_choices _lowerCamelCase = AlbertForMultipleChoice(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) lowercase__ : int = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) lowercase__ : str = True def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): _lowerCamelCase = super()._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) if return_labels: if model_class in get_values(lowerCamelCase__ ): _lowerCamelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowerCamelCase__ ) _lowerCamelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase__ ) return inputs_dict def snake_case__ ( self ): _lowerCamelCase = AlbertModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCamelCase = type self.model_tester.create_and_check_model(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = AlbertModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = AlbertModel.from_pretrained('''albert-base-v2''' ) _lowerCamelCase = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) _lowerCamelCase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ )[0] _lowerCamelCase = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCamelCase__ , atol=1e-4 ) )
623
"""simple docstring""" import unittest from transformers import BigBirdConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax from transformers.models.big_bird.modeling_flax_big_bird import ( FlaxBigBirdForCausalLM, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForPreTraining, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, FlaxBigBirdModel, ) class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=2 , lowerCamelCase__=5_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=2 , lowerCamelCase__=7 , lowerCamelCase__="gelu_new" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=4 , lowerCamelCase__="block_sparse" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=3 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_attention_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_choices _lowerCamelCase = rescale_embeddings _lowerCamelCase = attention_type _lowerCamelCase = use_bias _lowerCamelCase = block_size _lowerCamelCase = num_random_blocks def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_attention_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = BigBirdConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , attention_type=self.attention_type , block_size=self.block_size , num_random_blocks=self.num_random_blocks , use_bias=self.use_bias , rescale_embeddings=self.rescale_embeddings , ) return config, input_ids, token_type_ids, attention_mask def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask, } return config, inputs_dict @require_flax class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[str] = ( ( FlaxBigBirdForCausalLM, FlaxBigBirdModel, FlaxBigBirdForPreTraining, FlaxBigBirdForMaskedLM, FlaxBigBirdForMultipleChoice, FlaxBigBirdForQuestionAnswering, FlaxBigBirdForSequenceClassification, FlaxBigBirdForTokenClassification, ) if is_flax_available() else () ) lowercase__ : Any = False lowercase__ : Optional[int] = False def snake_case__ ( self ): _lowerCamelCase = FlaxBigBirdModelTester(self ) @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_save_pretrained() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_from_pretrained_with_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_no_automatic_init() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): super().test_hidden_states_output() @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''google/bigbird-roberta-base''' ) self.assertIsNotNone(lowerCamelCase__ ) def snake_case__ ( self ): if self.test_attn_probs: super().test_attention_outputs() @slow # copied from `test_modeling_flax_common` because it takes much longer than other models def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def model_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ , **lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = model_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=1e-5 , lowerCamelCase__="outputs" , lowerCamelCase__=None ): # `bigbird_block_sparse_attention` in `FlaxBigBird` returns `attention_probs = None`, while in PyTorch version, # an effort was done to return `attention_probs` (yet to be verified). if name.startswith('''outputs.attentions''' ): return else: super().check_pt_flax_outputs(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ )
623
1
"""simple docstring""" import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets __SCREAMING_SNAKE_CASE : int = datasets.logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' __SCREAMING_SNAKE_CASE : int = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : int , lowercase_ : Union[str, Any]=False , lowercase_ : Optional[int]=False , lowercase_ : List[Any]=True , lowercase_ : List[str]=False , lowercase_ : Tuple="dummy_doc" ) -> str: _lowerCamelCase = {doc: key_lines} _lowerCamelCase = {doc: sys_lines} _lowerCamelCase = {} _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase , _lowerCamelCase = reader.get_doc_mentions(lowercase_ , key_doc_lines[doc] , lowercase_ ) key_singletons_num += singletons_num if NP_only or min_span: _lowerCamelCase = reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) _lowerCamelCase , _lowerCamelCase = reader.get_doc_mentions(lowercase_ , sys_doc_lines[doc] , lowercase_ ) sys_singletons_num += singletons_num if NP_only or min_span: _lowerCamelCase = reader.set_annotated_parse_trees(lowercase_ , key_doc_lines[doc] , lowercase_ , lowercase_ ) if remove_nested: _lowerCamelCase , _lowerCamelCase = reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _lowerCamelCase , _lowerCamelCase = reader.remove_nested_coref_mentions(lowercase_ , lowercase_ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _lowerCamelCase = reader.get_mention_assignments(lowercase_ , lowercase_ ) _lowerCamelCase = reader.get_mention_assignments(lowercase_ , lowercase_ ) _lowerCamelCase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' F"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( '''Number of resulting singleton clusters in the key ''' F"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( F"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ '''files, respectively''' ) return doc_coref_infos def lowerCAmelCase_( lowercase_ : int , lowercase_ : str , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Dict , lowercase_ : Dict , lowercase_ : int ) -> str: _lowerCamelCase = get_coref_infos(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = {} _lowerCamelCase = 0 _lowerCamelCase = 0 for name, metric in metrics: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = evaluator.evaluate_documents(lowercase_ , lowercase_ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"""{name}/recall""": recall, F"""{name}/precision""": precision, F"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , F"""Recall: {recall * 1_00:.2f}""" , F""" Precision: {precision * 1_00:.2f}""" , F""" F1: {fa * 1_00:.2f}""" , ) if conll_subparts_num == 3: _lowerCamelCase = (conll / 3) * 1_00 logger.info(F"""CoNLL score: {conll:.2f}""" ) output_scores.update({'''conll_score''': conll} ) return output_scores def lowerCAmelCase_( lowercase_ : int ) -> int: _lowerCamelCase = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: _lowerCamelCase = line.split()[5] if not parse_col == "-": _lowerCamelCase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) , codebase_urls=['''https://github.com/ns-moosavi/coval'''] , reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=False ): _lowerCamelCase = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: _lowerCamelCase = util.check_gold_parse_annotation(lowerCamelCase__ ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _lowerCamelCase = evaluate( key_lines=lowerCamelCase__ , sys_lines=lowerCamelCase__ , metrics=lowerCamelCase__ , NP_only=lowerCamelCase__ , remove_nested=lowerCamelCase__ , keep_singletons=lowerCamelCase__ , min_span=lowerCamelCase__ , ) return score
623
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = StableDiffusionXLImgaImgPipeline lowercase__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} lowercase__ : int = PipelineTesterMixin.required_optional_params - {'latents'} lowercase__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS lowercase__ : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS lowercase__ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def snake_case__ ( self ): torch.manual_seed(0 ) _lowerCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , addition_embed_type='''text_time''' , addition_time_embed_dim=8 , transformer_layers_per_block=(1, 2) , projection_class_embeddings_input_dim=8_0 , cross_attention_dim=6_4 , ) _lowerCamelCase = EulerDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , steps_offset=1 , beta_schedule='''scaled_linear''' , timestep_spacing='''leading''' , ) torch.manual_seed(0 ) _lowerCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _lowerCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=3_2 , ) _lowerCamelCase = CLIPTextModel(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = CLIPTextModelWithProjection(lowerCamelCase__ ) _lowerCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' , local_files_only=lowerCamelCase__ ) _lowerCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''text_encoder_2''': text_encoder_a, '''tokenizer_2''': tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=0 ): _lowerCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) _lowerCamelCase = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): _lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 5.0, '''output_type''': '''numpy''', '''strength''': 0.7_5, } return inputs def snake_case__ ( self ): _lowerCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = sd_pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowerCamelCase = np.array([0.4_6_5_6, 0.4_8_4_0, 0.4_4_3_9, 0.6_6_9_8, 0.5_5_7_4, 0.4_5_2_4, 0.5_7_9_9, 0.5_9_4_3, 0.5_1_6_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case__ ( self ): super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case__ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.get_dummy_components() _lowerCamelCase = StableDiffusionXLImgaImgPipeline(**lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) _lowerCamelCase = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # forward without prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = negative_prompt _lowerCamelCase = 3 * [inputs['''prompt''']] _lowerCamelCase = sd_pipe(**lowerCamelCase__ ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # forward with prompt embeds _lowerCamelCase = self.get_dummy_inputs(lowerCamelCase__ ) _lowerCamelCase = 3 * ['''this is a negative prompt'''] _lowerCamelCase = 3 * [inputs.pop('''prompt''' )] ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = sd_pipe.encode_prompt(lowerCamelCase__ , negative_prompt=lowerCamelCase__ ) _lowerCamelCase = sd_pipe( **lowerCamelCase__ , prompt_embeds=lowerCamelCase__ , negative_prompt_embeds=lowerCamelCase__ , pooled_prompt_embeds=lowerCamelCase__ , negative_pooled_prompt_embeds=lowerCamelCase__ , ) _lowerCamelCase = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__="cpu" , lowerCamelCase__=torch.floataa , lowerCamelCase__=0 ): _lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _lowerCamelCase = np.random.RandomState(lowerCamelCase__ ).standard_normal((1, 4, 6_4, 6_4) ) _lowerCamelCase = torch.from_numpy(lowerCamelCase__ ).to(device=lowerCamelCase__ , dtype=lowerCamelCase__ ) _lowerCamelCase = { '''prompt''': '''a photograph of an astronaut riding a horse''', '''latents''': latents, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self ): _lowerCamelCase = DiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-base''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = self.get_inputs(lowerCamelCase__ ) _lowerCamelCase = pipe(**lowerCamelCase__ ).images _lowerCamelCase = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.4_9_4_9_3, 0.4_7_8_9_6, 0.4_0_7_9_8, 0.5_4_2_1_4, 0.5_3_2_1_2, 0.4_8_2_0_2, 0.4_7_6_5_6, 0.4_6_3_2_9, 0.4_8_5_0_6] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
623
1
"""simple docstring""" # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = subprocess.check_output('''git merge-base main HEAD'''.split()).decode('''utf-8''') __SCREAMING_SNAKE_CASE : List[str] = subprocess.check_output(F"""git diff --name-only {fork_point_sha}""".split()).decode('''utf-8''').split() __SCREAMING_SNAKE_CASE : List[Any] = '''|'''.join(sys.argv[1:]) __SCREAMING_SNAKE_CASE : Optional[int] = re.compile(RF"""^({joined_dirs}).*?\.py$""") __SCREAMING_SNAKE_CASE : List[str] = [x for x in modified_files if regex.match(x)] print(''' '''.join(relevant_modified_files), end='''''')
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_xlm''': ['''XLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMConfig''', '''XLMOnnxConfig'''], '''tokenization_xlm''': ['''XLMTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMForMultipleChoice''', '''XLMForQuestionAnswering''', '''XLMForQuestionAnsweringSimple''', '''XLMForSequenceClassification''', '''XLMForTokenClassification''', '''XLMModel''', '''XLMPreTrainedModel''', '''XLMWithLMHeadModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Union[str, Any] = [ '''TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMForMultipleChoice''', '''TFXLMForQuestionAnsweringSimple''', '''TFXLMForSequenceClassification''', '''TFXLMForTokenClassification''', '''TFXLMMainLayer''', '''TFXLMModel''', '''TFXLMPreTrainedModel''', '''TFXLMWithLMHeadModel''', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __SCREAMING_SNAKE_CASE : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowerCAmelCase_( lowercase_ : str = "laptop" ) -> DataFrame: _lowerCamelCase = F"""https://www.amazon.in/laptop/s?k={product}""" _lowerCamelCase = { '''User-Agent''': '''Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36''', '''Accept-Language''': '''en-US, en;q=0.5''', } _lowerCamelCase = BeautifulSoup(requests.get(lowercase_ , headers=lowercase_ ).text ) # Initialize a Pandas dataframe with the column titles _lowerCamelCase = DataFrame( columns=[ '''Product Title''', '''Product Link''', '''Current Price of the product''', '''Product Rating''', '''MRP of the product''', '''Discount''', ] ) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( '''div''' , attrs={'''class''': '''s-result-item''', '''data-component-type''': '''s-search-result'''} , ) , soup.find_all('''div''' , attrs={'''class''': '''a-row a-size-base a-color-base'''} ) , ): try: _lowerCamelCase = item.ha.text _lowerCamelCase = '''https://www.amazon.in/''' + item.ha.a['''href'''] _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-offscreen'''} ).text try: _lowerCamelCase = item.find('''span''' , attrs={'''class''': '''a-icon-alt'''} ).text except AttributeError: _lowerCamelCase = '''Not available''' try: _lowerCamelCase = ( '''₹''' + item.find( '''span''' , attrs={'''class''': '''a-price a-text-price'''} ).text.split('''₹''' )[1] ) except AttributeError: _lowerCamelCase = '''''' try: _lowerCamelCase = float( ( ( float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) - float(product_price.strip('''₹''' ).replace(''',''' , '''''' ) ) ) / float(product_mrp.strip('''₹''' ).replace(''',''' , '''''' ) ) ) * 1_00 ) except ValueError: _lowerCamelCase = float('''nan''' ) except AttributeError: pass _lowerCamelCase = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] _lowerCamelCase = ''' ''' _lowerCamelCase = ''' ''' data_frame.index += 1 return data_frame if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = '''headphones''' get_amazon_product_data(product).to_csv(F"""Amazon Product Data for {product}.csv""")
623
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Dict = random.Random() def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : int=1.0 , lowercase_ : str=None , lowercase_ : Optional[int]=None ) -> Any: if rng is None: _lowerCamelCase = global_rng _lowerCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=4_0_0 , lowerCamelCase__=2_0_0_0 , lowerCamelCase__=1_0 , lowerCamelCase__=1_6_0 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=4_0_0_0 , lowerCamelCase__=False , lowerCamelCase__=True , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = min_seq_length _lowerCamelCase = max_seq_length _lowerCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) _lowerCamelCase = padding_value _lowerCamelCase = sampling_rate _lowerCamelCase = return_attention_mask _lowerCamelCase = do_normalize _lowerCamelCase = feature_size _lowerCamelCase = chunk_length _lowerCamelCase = hop_length def snake_case__ ( self ): return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case__ ( self , lowerCamelCase__=False , lowerCamelCase__=False ): def _flatten(lowerCamelCase__ ): return list(itertools.chain(*lowerCamelCase__ ) ) if equal_length: _lowerCamelCase = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size _lowerCamelCase = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = WhisperFeatureExtractor if is_speech_available() else None def snake_case__ ( self ): _lowerCamelCase = WhisperFeatureExtractionTester(self ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = feat_extract_first.save_pretrained(lowerCamelCase__ )[0] check_json_file_has_correct_format(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase = os.path.join(lowerCamelCase__ , '''feat_extract.json''' ) feat_extract_first.to_json_file(lowerCamelCase__ ) _lowerCamelCase = self.feature_extraction_class.from_json_file(lowerCamelCase__ ) _lowerCamelCase = feat_extract_first.to_dict() _lowerCamelCase = feat_extract_second.to_dict() _lowerCamelCase = feat_extract_first.mel_filters _lowerCamelCase = feat_extract_second.mel_filters self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ ) ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 _lowerCamelCase = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] # Test feature size _lowerCamelCase = feature_extractor(lowerCamelCase__ , padding='''max_length''' , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input _lowerCamelCase = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test batched _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. _lowerCamelCase = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] _lowerCamelCase = np.asarray(lowerCamelCase__ ) _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) # Test truncation required _lowerCamelCase = [floats_list((1, x) )[0] for x in range(2_0_0 , (feature_extractor.n_samples + 5_0_0) , 2_0_0 )] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs] _lowerCamelCase = [x[: feature_extractor.n_samples] for x in speech_inputs] _lowerCamelCase = [np.asarray(lowerCamelCase__ ) for speech_input in speech_inputs_truncated] _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertTrue(np.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-3 ) ) def snake_case__ ( self ): import torch _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) _lowerCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) _lowerCamelCase = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech _lowerCamelCase = ds.sort('''id''' ).select(range(lowerCamelCase__ ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def snake_case__ ( self ): # fmt: off _lowerCamelCase = torch.tensor( [ 0.1_1_9_3, -0.0_9_4_6, -0.1_0_9_8, -0.0_1_9_6, 0.0_2_2_5, -0.0_6_9_0, -0.1_7_3_6, 0.0_9_5_1, 0.0_9_7_1, -0.0_8_1_7, -0.0_7_0_2, 0.0_1_6_2, 0.0_2_6_0, 0.0_0_1_7, -0.0_1_9_2, -0.1_6_7_8, 0.0_7_0_9, -0.1_8_6_7, -0.0_6_5_5, -0.0_2_7_4, -0.0_2_3_4, -0.1_8_8_4, -0.0_5_1_6, -0.0_5_5_4, -0.0_2_7_4, -0.1_4_2_5, -0.1_4_2_3, 0.0_8_3_7, 0.0_3_7_7, -0.0_8_5_4 ] ) # fmt: on _lowerCamelCase = self._load_datasamples(1 ) _lowerCamelCase = WhisperFeatureExtractor() _lowerCamelCase = feature_extractor(lowerCamelCase__ , return_tensors='''pt''' ).input_features self.assertEqual(input_features.shape , (1, 8_0, 3_0_0_0) ) self.assertTrue(torch.allclose(input_features[0, 0, :3_0] , lowerCamelCase__ , atol=1e-4 ) ) def snake_case__ ( self ): _lowerCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) _lowerCamelCase = self._load_datasamples(1 )[0] _lowerCamelCase = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5_5_3_5 # Rescale to [0, 65535] to show issue _lowerCamelCase = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=lowerCamelCase__ )[0] self.assertTrue(np.all(np.mean(lowerCamelCase__ ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(lowerCamelCase__ ) - 1 ) < 1e-3 ) )
623
1
"""simple docstring""" import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html __SCREAMING_SNAKE_CASE : Optional[Any] = '''platform''' import jax import jax.numpy as jnp from transformers.models.blenderbot_small.modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, shift_tokens_right, ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : int , lowercase_ : int=None , lowercase_ : List[Any]=None , lowercase_ : Dict=None , lowercase_ : Union[str, Any]=None , lowercase_ : Any=None , lowercase_ : List[str]=None , ) -> Optional[int]: if attention_mask is None: _lowerCamelCase = np.where(input_ids != config.pad_token_id , 1 , 0 ) if decoder_attention_mask is None: _lowerCamelCase = np.where(decoder_input_ids != config.pad_token_id , 1 , 0 ) if head_mask is None: _lowerCamelCase = np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: _lowerCamelCase = np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: _lowerCamelCase = np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=9_9 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=0.0_2 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = eos_token_id _lowerCamelCase = pad_token_id _lowerCamelCase = bos_token_id _lowerCamelCase = initializer_range def snake_case__ ( self ): _lowerCamelCase = np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) _lowerCamelCase = np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) _lowerCamelCase = shift_tokens_right(lowerCamelCase__ , 1 , 2 ) _lowerCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCamelCase__ , ) _lowerCamelCase = prepare_blenderbot_inputs_dict(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return config, inputs_dict def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.prepare_config_and_inputs() return config, inputs_dict def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = 2_0 _lowerCamelCase = model_class_name(lowerCamelCase__ ) _lowerCamelCase = model.encode(inputs_dict['''input_ids'''] ) _lowerCamelCase , _lowerCamelCase = ( inputs_dict['''decoder_input_ids'''], inputs_dict['''decoder_attention_mask'''], ) _lowerCamelCase = model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='''i4''' ) _lowerCamelCase = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _lowerCamelCase = model.decode( decoder_input_ids[:, :-1] , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , past_key_values=lowerCamelCase__ , decoder_position_ids=lowerCamelCase__ , ) _lowerCamelCase = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='''i4''' ) _lowerCamelCase = model.decode( decoder_input_ids[:, -1:] , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCamelCase__ , ) _lowerCamelCase = model.decode(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F"""Max diff is {diff}""" ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = 2_0 _lowerCamelCase = model_class_name(lowerCamelCase__ ) _lowerCamelCase = model.encode(inputs_dict['''input_ids'''] ) _lowerCamelCase , _lowerCamelCase = ( inputs_dict['''decoder_input_ids'''], inputs_dict['''decoder_attention_mask'''], ) _lowerCamelCase = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) _lowerCamelCase = model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) _lowerCamelCase = model.decode( decoder_input_ids[:, :-1] , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , past_key_values=lowerCamelCase__ , decoder_position_ids=lowerCamelCase__ , ) _lowerCamelCase = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='''i4''' ) _lowerCamelCase = model.decode( decoder_input_ids[:, -1:] , lowerCamelCase__ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCamelCase__ , decoder_position_ids=lowerCamelCase__ , ) _lowerCamelCase = model.decode(lowerCamelCase__ , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ ) _lowerCamelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1e-3 , msg=F"""Max diff is {diff}""" ) @require_flax class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = 99 def snake_case__ ( self ): _lowerCamelCase = np.array( [ [7_1, 8_2, 1_8, 3_3, 4_6, 9_1, 2], [6_8, 3_4, 2_6, 5_8, 3_0, 8_2, 2], [5, 9_7, 1_7, 3_9, 9_4, 4_0, 2], [7_6, 8_3, 9_4, 2_5, 7_0, 7_8, 2], [8_7, 5_9, 4_1, 3_5, 4_8, 6_6, 2], [5_5, 1_3, 1_6, 5_8, 5, 2, 1], # note padding [6_4, 2_7, 3_1, 5_1, 1_2, 7_5, 2], [5_2, 6_4, 8_6, 1_7, 8_3, 3_9, 2], [4_8, 6_1, 9, 2_4, 7_1, 8_2, 2], [2_6, 1, 6_0, 4_8, 2_2, 1_3, 2], [2_1, 5, 6_2, 2_8, 1_4, 7_6, 2], [4_5, 9_8, 3_7, 8_6, 5_9, 4_8, 2], [7_0, 7_0, 5_0, 9, 2_8, 0, 2], ] , dtype=np.intaa , ) _lowerCamelCase = input_ids.shape[0] _lowerCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=2_4 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=3_2 , decoder_ffn_dim=3_2 , max_position_embeddings=4_8 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self._get_config_and_data() _lowerCamelCase = FlaxBlenderbotSmallForConditionalGeneration(lowerCamelCase__ ) _lowerCamelCase = lm_model(input_ids=lowerCamelCase__ ) _lowerCamelCase = (batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs['''logits'''].shape , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=1_4 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=4_8 , ) _lowerCamelCase = FlaxBlenderbotSmallForConditionalGeneration(lowerCamelCase__ ) _lowerCamelCase = np.array([[7_1, 8_2, 1_8, 3_3, 4_6, 9_1, 2], [6_8, 3_4, 2_6, 5_8, 3_0, 2, 1]] , dtype=np.intaa ) _lowerCamelCase = np.array([[8_2, 7_1, 8_2, 1_8, 2], [5_8, 6_8, 2, 1, 1]] , dtype=np.intaa ) _lowerCamelCase = lm_model(input_ids=lowerCamelCase__ , decoder_input_ids=lowerCamelCase__ ) _lowerCamelCase = (*summary.shape, config.vocab_size) self.assertEqual(outputs['''logits'''].shape , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = np.array([[7_1, 8_2, 1_8, 3_3, 2, 1, 1], [6_8, 3_4, 2_6, 5_8, 3_0, 8_2, 2]] , dtype=np.intaa ) _lowerCamelCase = shift_tokens_right(lowerCamelCase__ , 1 , 2 ) _lowerCamelCase = np.equal(lowerCamelCase__ , 1 ).astype(np.floataa ).sum() _lowerCamelCase = np.equal(lowerCamelCase__ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCamelCase__ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class lowerCamelCase_( A__, unittest.TestCase, A__ ): '''simple docstring''' lowercase__ : Optional[int] = True lowercase__ : Optional[Any] = ( ( FlaxBlenderbotSmallModel, FlaxBlenderbotSmallForConditionalGeneration, ) if is_flax_available() else () ) lowercase__ : str = (FlaxBlenderbotSmallForConditionalGeneration,) if is_flax_available() else () def snake_case__ ( self ): _lowerCamelCase = FlaxBlenderbotSmallModelTester(self ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model_class(lowerCamelCase__ ) @jax.jit def encode_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model.encode(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = encode_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = encode_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = model.encode(inputs_dict['''input_ids'''] , inputs_dict['''attention_mask'''] ) _lowerCamelCase = { '''decoder_input_ids''': inputs_dict['''decoder_input_ids'''], '''decoder_attention_mask''': inputs_dict['''decoder_attention_mask'''], '''encoder_outputs''': encoder_outputs, } @jax.jit def decode_jitted(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return model.decode( decoder_input_ids=lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , encoder_outputs=lowerCamelCase__ , ) with self.subTest('''JIT Enabled''' ): _lowerCamelCase = decode_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): _lowerCamelCase = decode_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def snake_case__ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase = model_class_name.from_pretrained('''facebook/blenderbot_small-90M''' ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids _lowerCamelCase = np.ones((1, 1) ) * model.config.eos_token_id _lowerCamelCase = model(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
623
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] _lowerCamelCase = True for i in range(lowercase_ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: _lowerCamelCase = True if a[i].islower(): _lowerCamelCase = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: __SCREAMING_SNAKE_CASE : Any = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=3 , lowerCamelCase__=1_8 , lowerCamelCase__=3_0 , lowerCamelCase__=4_0_0 , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=None , ): _lowerCamelCase = size if size is not None else {'''height''': 2_0, '''width''': 2_0} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = image_size _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = do_convert_rgb _lowerCamelCase = [5_1_2, 1_0_2_4, 2_0_4_8, 4_0_9_6] _lowerCamelCase = patch_size if patch_size is not None else {'''height''': 1_6, '''width''': 1_6} def snake_case__ ( self ): return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def snake_case__ ( self ): _lowerCamelCase = '''https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg''' _lowerCamelCase = Image.open(requests.get(lowerCamelCase__ , stream=lowerCamelCase__ ).raw ).convert('''RGB''' ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11, reason='`Pix2StructImageProcessor` requires `torch>=1.11.0`.', ) @require_torch @require_vision class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = PixaStructImageProcessor if is_vision_available() else None def snake_case__ ( self ): _lowerCamelCase = PixaStructImageProcessingTester(self ) @property def snake_case__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self ): _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_convert_rgb''' ) ) def snake_case__ ( self ): _lowerCamelCase = self.image_processor_tester.prepare_dummy_image() _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) _lowerCamelCase = 2_0_4_8 _lowerCamelCase = image_processor(lowerCamelCase__ , return_tensors='''pt''' , max_patches=lowerCamelCase__ ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() , torch.tensor(0.0_6_0_6 ) , atol=1e-3 , rtol=1e-3 ) ) def snake_case__ ( self ): # Initialize image_processor _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input _lowerCamelCase = ( (self.image_processor_tester.patch_size['''height'''] * self.image_processor_tester.patch_size['''width''']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _lowerCamelCase = image_processor( image_inputs[0] , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _lowerCamelCase = image_processor( lowerCamelCase__ , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def snake_case__ ( self ): # Initialize image_processor _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input _lowerCamelCase = ( (self.image_processor_tester.patch_size['''height'''] * self.image_processor_tester.patch_size['''width''']) * self.image_processor_tester.num_channels ) + 2 _lowerCamelCase = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(lowerCamelCase__ ): _lowerCamelCase = image_processor( image_inputs[0] , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches _lowerCamelCase = '''Hello''' _lowerCamelCase = image_processor( image_inputs[0] , return_tensors='''pt''' , max_patches=lowerCamelCase__ , header_text=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _lowerCamelCase = image_processor( lowerCamelCase__ , return_tensors='''pt''' , max_patches=lowerCamelCase__ , header_text=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def snake_case__ ( self ): # Initialize image_processor _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , np.ndarray ) _lowerCamelCase = ( (self.image_processor_tester.patch_size['''height'''] * self.image_processor_tester.patch_size['''width''']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _lowerCamelCase = image_processor( image_inputs[0] , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _lowerCamelCase = image_processor( lowerCamelCase__ , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) def snake_case__ ( self ): # Initialize image_processor _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , torch.Tensor ) # Test not batched input _lowerCamelCase = ( (self.image_processor_tester.patch_size['''height'''] * self.image_processor_tester.patch_size['''width''']) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _lowerCamelCase = image_processor( image_inputs[0] , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _lowerCamelCase = image_processor( lowerCamelCase__ , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , ) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11, reason='`Pix2StructImageProcessor` requires `torch>=1.11.0`.', ) @require_torch @require_vision class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = PixaStructImageProcessor if is_vision_available() else None def snake_case__ ( self ): _lowerCamelCase = PixaStructImageProcessingTester(self , num_channels=4 ) _lowerCamelCase = 3 @property def snake_case__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self ): _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_convert_rgb''' ) ) def snake_case__ ( self ): # Initialize image_processor _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input _lowerCamelCase = ( (self.image_processor_tester.patch_size['''height'''] * self.image_processor_tester.patch_size['''width''']) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input _lowerCamelCase = image_processor( image_inputs[0] , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (1, max_patch, expected_hidden_dim) , ) # Test batched _lowerCamelCase = image_processor( lowerCamelCase__ , return_tensors='''pt''' , max_patches=lowerCamelCase__ ).flattened_patches self.assertEqual( encoded_images.shape , (self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) , )
623
"""simple docstring""" import numpy as np def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def lowerCAmelCase_( lowercase_ : np.array ) -> np.array: return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Dict = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : str = BartphoTokenizer lowercase__ : Union[str, Any] = False lowercase__ : List[Any] = True def snake_case__ ( self ): super().setUp() _lowerCamelCase = ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {'''unk_token''': '''<unk>'''} _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''monolingual_vocab_file'''] ) with open(self.monolingual_vocab_file , '''w''' , encoding='''utf-8''' ) as fp: for token in vocab_tokens: fp.write(F"""{token} {vocab_tokens[token]}\n""" ) _lowerCamelCase = BartphoTokenizer(lowerCamelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self , **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = '''This is a là test''' _lowerCamelCase = '''This is a<unk><unk> test''' return input_text, output_text def snake_case__ ( self ): _lowerCamelCase = BartphoTokenizer(lowerCamelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) _lowerCamelCase = '''This is a là test''' _lowerCamelCase = '''▁This ▁is ▁a ▁l à ▁t est'''.split() _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokens + [tokenizer.unk_token] _lowerCamelCase = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , lowerCamelCase__ )
623
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''configuration_vision_encoder_decoder''': ['''VisionEncoderDecoderConfig''', '''VisionEncoderDecoderOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = ['''VisionEncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = ['''TFVisionEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''FlaxVisionEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
623
1
"""simple docstring""" import warnings warnings.warn( '''memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: ''' '''`from accelerate import find_executable_batch_size` to avoid this warning.''', FutureWarning, )
623
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float ) -> dict[str, float]: if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(lowercase_ , 2 ) - pow(lowercase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(lowercase_ , 2 ) + pow(lowercase_ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
623
1
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __SCREAMING_SNAKE_CASE : int = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') require_version('''datasets>=1.14.0''', '''To fix: pip install -r examples/pytorch/audio-classification/requirements.txt''') def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : float , lowercase_ : int = 1_60_00 ) -> Tuple: _lowerCamelCase = int(round(sample_rate * max_length ) ) if len(lowercase_ ) <= sample_length: return wav _lowerCamelCase = randint(0 , len(lowercase_ ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : Optional[str] = field(default=A__, metadata={'help': 'Name of a dataset from the datasets package'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'A file containing the training audio paths and labels.'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'A file containing the validation audio paths and labels.'} ) lowercase__ : str = field( default='train', metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' }, ) lowercase__ : str = field( default='validation', metadata={ 'help': ( 'The name of the training data set split to use (via the datasets library). Defaults to \'validation\'' ) }, ) lowercase__ : str = field( default='audio', metadata={'help': 'The name of the dataset column containing the audio data. Defaults to \'audio\''}, ) lowercase__ : str = field( default='label', metadata={'help': 'The name of the dataset column containing the labels. Defaults to \'label\''} ) lowercase__ : Optional[int] = field( default=A__, metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) }, ) lowercase__ : Optional[int] = field( default=A__, metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) }, ) lowercase__ : float = field( default=20, metadata={'help': 'Audio clips will be randomly cut to this length during training if the value is set.'}, ) @dataclass class lowerCamelCase_: '''simple docstring''' lowercase__ : str = field( default='facebook/wav2vec2-base', metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'}, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Where do you want to store the pretrained models downloaded from the Hub'} ) lowercase__ : str = field( default='main', metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'}, ) lowercase__ : Optional[str] = field( default=A__, metadata={'help': 'Name or path of preprocessor config.'} ) lowercase__ : bool = field( default=A__, metadata={'help': 'Whether to freeze the feature encoder layers of the model.'} ) lowercase__ : bool = field( default=A__, metadata={'help': 'Whether to generate an attention mask in the feature extractor.'} ) lowercase__ : bool = field( default=A__, metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) }, ) lowercase__ : Optional[bool] = field( default=A__, metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) lowercase__ : bool = field( default=A__, metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'}, ) def snake_case__ ( self ): if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( '''The argument `--freeze_feature_extractor` is deprecated and ''' '''will be removed in a future version. Use `--freeze_feature_encoder`''' '''instead. Setting `freeze_feature_encoder==True`.''' , lowerCamelCase__ , ) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( '''The argument `--freeze_feature_extractor` is deprecated and ''' '''should not be used in combination with `--freeze_feature_encoder`.''' '''Only make use of `--freeze_feature_encoder`.''' ) def lowerCAmelCase_( ) -> int: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowerCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_audio_classification''' , lowercase_ , lowercase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowerCamelCase = training_args.get_process_log_level() logger.setLevel(lowercase_ ) transformers.utils.logging.set_verbosity(lowercase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} """ + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. _lowerCamelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCamelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to train from scratch.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset and prepare it for the audio classification task. _lowerCamelCase = DatasetDict() _lowerCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( F"""--audio_column_name {data_args.audio_column_name} not found in dataset '{data_args.dataset_name}'. """ '''Make sure to set `--audio_column_name` to the correct audio column - one of ''' F"""{", ".join(raw_datasets["train"].column_names )}.""" ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( F"""--label_column_name {data_args.label_column_name} not found in dataset '{data_args.dataset_name}'. """ '''Make sure to set `--label_column_name` to the correct text column - one of ''' F"""{", ".join(raw_datasets["train"].column_names )}.""" ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy _lowerCamelCase = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. _lowerCamelCase = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) _lowerCamelCase = feature_extractor.model_input_names[0] def train_transforms(lowercase_ : Dict ): _lowerCamelCase = [] for audio in batch[data_args.audio_column_name]: _lowerCamelCase = random_subsample( audio['''array'''] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(lowercase_ ) _lowerCamelCase = feature_extractor(lowercase_ , sampling_rate=feature_extractor.sampling_rate ) _lowerCamelCase = {model_input_name: inputs.get(lowercase_ )} _lowerCamelCase = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(lowercase_ : str ): _lowerCamelCase = [audio['''array'''] for audio in batch[data_args.audio_column_name]] _lowerCamelCase = feature_extractor(lowercase_ , sampling_rate=feature_extractor.sampling_rate ) _lowerCamelCase = {model_input_name: inputs.get(lowercase_ )} _lowerCamelCase = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _lowerCamelCase = raw_datasets['''train'''].features[data_args.label_column_name].names _lowerCamelCase , _lowerCamelCase = {}, {} for i, label in enumerate(lowercase_ ): _lowerCamelCase = str(lowercase_ ) _lowerCamelCase = label # Load the accuracy metric from the datasets package _lowerCamelCase = evaluate.load('''accuracy''' ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(lowercase_ : Optional[int] ): _lowerCamelCase = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=lowercase_ , references=eval_pred.label_ids ) _lowerCamelCase = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(lowercase_ ) , labelaid=lowercase_ , idalabel=lowercase_ , finetuning_task='''audio-classification''' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _lowerCamelCase = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowercase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: _lowerCamelCase = ( raw_datasets['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(lowercase_ , output_all_columns=lowercase_ ) if training_args.do_eval: if data_args.max_eval_samples is not None: _lowerCamelCase = ( raw_datasets['''eval'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(lowercase_ , output_all_columns=lowercase_ ) # Initialize our trainer _lowerCamelCase = Trainer( model=lowercase_ , args=lowercase_ , train_dataset=raw_datasets['''train'''] if training_args.do_train else None , eval_dataset=raw_datasets['''eval'''] if training_args.do_eval else None , compute_metrics=lowercase_ , tokenizer=lowercase_ , ) # Training if training_args.do_train: _lowerCamelCase = None if training_args.resume_from_checkpoint is not None: _lowerCamelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCamelCase = last_checkpoint _lowerCamelCase = trainer.train(resume_from_checkpoint=lowercase_ ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _lowerCamelCase = trainer.evaluate() trainer.log_metrics('''eval''' , lowercase_ ) trainer.save_metrics('''eval''' , lowercase_ ) # Write model card and (optionally) push to hub _lowerCamelCase = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''audio-classification''', '''dataset''': data_args.dataset_name, '''tags''': ['''audio-classification'''], } if training_args.push_to_hub: trainer.push_to_hub(**lowercase_ ) else: trainer.create_model_card(**lowercase_ ) if __name__ == "__main__": main()
623
"""simple docstring""" from __future__ import annotations from typing import Any def lowerCAmelCase_( lowercase_ : list[Any] ) -> None: create_state_space_tree(lowercase_ , [] , 0 ) def lowerCAmelCase_( lowercase_ : list[Any] , lowercase_ : list[Any] , lowercase_ : int ) -> None: if index == len(lowercase_ ): print(lowercase_ ) return create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(lowercase_ , lowercase_ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __SCREAMING_SNAKE_CASE : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['''A''', '''B''', '''C''']) generate_all_subsequences(seq)
623
1