code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version SCREAMING_SNAKE_CASE__:Optional[int] = logging.getLogger(__name__) require_version("""pytorch_lightning>=1.0.4""") SCREAMING_SNAKE_CASE__:Optional[Any] = { """base""": AutoModel, """sequence-classification""": AutoModelForSequenceClassification, """question-answering""": AutoModelForQuestionAnswering, """pretraining""": AutoModelForPreTraining, """token-classification""": AutoModelForTokenClassification, """language-modeling""": AutoModelWithLMHead, """summarization""": AutoModelForSeqaSeqLM, """translation""": AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization SCREAMING_SNAKE_CASE__:Tuple = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } SCREAMING_SNAKE_CASE__:List[Any] = sorted(arg_to_scheduler.keys()) SCREAMING_SNAKE_CASE__:str = """{""" + """, """.join(arg_to_scheduler_choices) + """}""" class snake_case__ ( pl.LightningModule ): def __init__( self , lowerCamelCase , lowerCamelCase=None , lowerCamelCase="base" , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , **lowerCamelCase , ): super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(lowerCAmelCase__ ) __a = 0 __a = Path(self.hparams.output_dir ) __a = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: __a = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path , **({"num_labels": num_labels} if num_labels is not None else {}) , cache_dir=lowerCAmelCase__ , **lowerCAmelCase__ , ) else: __a = config __a = ("encoder_layerdrop", "decoder_layerdrop", "dropout", "attention_dropout") for p in extra_model_params: if getattr(self.hparams , lowerCAmelCase__ , lowerCAmelCase__ ): assert hasattr(self.config , lowerCAmelCase__ ), F"model config doesn't have a `{p}` attribute" setattr(self.config , lowerCAmelCase__ , getattr(self.hparams , lowerCAmelCase__ ) ) if tokenizer is None: __a = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path , cache_dir=lowerCAmelCase__ , ) else: __a = tokenizer __a = MODEL_MODES[mode] if model is None: __a = self.model_type.from_pretrained( self.hparams.model_name_or_path , from_tf=bool(".ckpt" in self.hparams.model_name_or_path ) , config=self.config , cache_dir=lowerCAmelCase__ , ) else: __a = model def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = self.model_type.from_pretrained(*lowerCAmelCase__ , **lowerCAmelCase__ ) def a__ ( self ): __a = arg_to_scheduler[self.hparams.lr_scheduler] __a = get_schedule_func( self.opt , num_warmup_steps=self.hparams.warmup_steps , num_training_steps=self.total_steps() ) __a = {"scheduler": scheduler, "interval": "step", "frequency": 1} return scheduler def a__ ( self ): __a = self.model __a = ["bias", "LayerNorm.weight"] __a = [ { "params": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay ) ], # check this named paramters "weight_decay": self.hparams.weight_decay, }, { "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay )], "weight_decay": 0.0, }, ] if self.hparams.adafactor: __a = Adafactor( lowerCAmelCase__ , lr=self.hparams.learning_rate , scale_parameter=lowerCAmelCase__ , relative_step=lowerCAmelCase__ ) else: __a = AdamW( lowerCAmelCase__ , lr=self.hparams.learning_rate , eps=self.hparams.adam_epsilon ) __a = optimizer __a = self.get_lr_scheduler() return [optimizer], [scheduler] def a__ ( self , lowerCamelCase , lowerCamelCase ): return self.validation_step(lowerCAmelCase__ , lowerCAmelCase__ ) def a__ ( self , lowerCamelCase ): return self.validation_end(lowerCAmelCase__ ) def a__ ( self ): __a = max(1 , self.hparams.gpus ) # TODO: consider num_tpu_cores __a = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def a__ ( self , lowerCamelCase ): if stage == "test": __a = len(self.test_dataloader().dataset ) else: __a = self.get_dataloader("train" , self.hparams.train_batch_size , shuffle=lowerCAmelCase__ ) __a = len(self.train_dataloader().dataset ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = False ): raise NotImplementedError("You must implement this for your task" ) def a__ ( self ): return self.train_loader def a__ ( self ): return self.get_dataloader("dev" , self.hparams.eval_batch_size , shuffle=lowerCAmelCase__ ) def a__ ( self ): return self.get_dataloader("test" , self.hparams.eval_batch_size , shuffle=lowerCAmelCase__ ) def a__ ( self , lowerCamelCase ): return os.path.join( self.hparams.data_dir , "cached_{}_{}_{}".format( lowerCAmelCase__ , list(filter(lowerCAmelCase__ , self.hparams.model_name_or_path.split("/" ) ) ).pop() , str(self.hparams.max_seq_length ) , ) , ) @pl.utilities.rank_zero_only def a__ ( self , lowerCamelCase ): __a = self.output_dir.joinpath("best_tfmr" ) __a = self.step_count self.model.save_pretrained(lowerCAmelCase__ ) self.tokenizer.save_pretrained(lowerCAmelCase__ ) @staticmethod def a__ ( lowerCamelCase , lowerCamelCase ): parser.add_argument( "--model_name_or_path" , default=lowerCAmelCase__ , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--config_name" , default="" , type=lowerCAmelCase__ , help="Pretrained config name or path if not the same as model_name" ) parser.add_argument( "--tokenizer_name" , default=lowerCAmelCase__ , type=lowerCAmelCase__ , help="Pretrained tokenizer name or path if not the same as model_name" , ) parser.add_argument( "--cache_dir" , default=str(Path(lowerCAmelCase__ ).parent / "test_run" / "cache" ) , type=lowerCAmelCase__ , help="Where do you want to store the pre-trained models downloaded from huggingface.co" , ) parser.add_argument( "--encoder_layerdrop" , type=lowerCAmelCase__ , help="Encoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--decoder_layerdrop" , type=lowerCAmelCase__ , help="Decoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--dropout" , type=lowerCAmelCase__ , help="Dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--attention_dropout" , type=lowerCAmelCase__ , help="Attention dropout probability (Optional). Goes into model.config" , ) parser.add_argument("--learning_rate" , default=5E-5 , type=lowerCAmelCase__ , help="The initial learning rate for Adam." ) parser.add_argument( "--lr_scheduler" , default="linear" , choices=lowerCAmelCase__ , metavar=lowerCAmelCase__ , type=lowerCAmelCase__ , help="Learning rate scheduler" , ) parser.add_argument("--weight_decay" , default=0.0 , type=lowerCAmelCase__ , help="Weight decay if we apply some." ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=lowerCAmelCase__ , help="Epsilon for Adam optimizer." ) parser.add_argument("--warmup_steps" , default=0 , type=lowerCAmelCase__ , help="Linear warmup over warmup_steps." ) parser.add_argument("--num_workers" , default=4 , type=lowerCAmelCase__ , help="kwarg passed to DataLoader" ) parser.add_argument("--num_train_epochs" , dest="max_epochs" , default=3 , type=lowerCAmelCase__ ) parser.add_argument("--train_batch_size" , default=32 , type=lowerCAmelCase__ ) parser.add_argument("--eval_batch_size" , default=32 , type=lowerCAmelCase__ ) parser.add_argument("--adafactor" , action="store_true" ) class snake_case__ ( pl.Callback ): def a__ ( self , lowerCamelCase , lowerCamelCase ): if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class snake_case__ ( pl.Callback ): def a__ ( self , lowerCamelCase , lowerCamelCase ): # print(pl_module.model.rag) for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(lowerCAmelCase__ ) class snake_case__ ( pl.Callback ): def a__ ( self , lowerCamelCase , lowerCamelCase ): __a = trainer.lr_schedulers[0]["scheduler"] __a = {F"lr_group_{i}": lr for i, lr in enumerate(lr_scheduler.get_lr() )} pl_module.logger.log_metrics(lowerCAmelCase__ ) def a__ ( self , lowerCamelCase , lowerCamelCase ): rank_zero_info("***** Validation results *****" ) __a = trainer.callback_metrics # Log results for key in sorted(lowerCAmelCase__ ): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(lowerCAmelCase__ , str(metrics[key] ) ) ) def a__ ( self , lowerCamelCase , lowerCamelCase ): rank_zero_info("***** Test results *****" ) __a = trainer.callback_metrics # Log and save results to file __a = os.path.join(pl_module.hparams.output_dir , "test_results.txt" ) with open(lowerCAmelCase__ , "w" ) as writer: for key in sorted(lowerCAmelCase__ ): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(lowerCAmelCase__ , str(metrics[key] ) ) ) writer.write("{} = {}\n".format(lowerCAmelCase__ , str(metrics[key] ) ) ) def _lowerCamelCase( a , a ): # To allow all pl args uncomment the following line # parser = pl.Trainer.add_argparse_args(parser) parser.add_argument( "--output_dir" , default=str(Path(__A ).parent / "test_run" / "model_checkpoints" ) , type=__A , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument( "--fp16" , action="store_true" , help="Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit" , ) parser.add_argument( "--fp16_opt_level" , type=__A , default="O2" , help=( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']." "See details at https://nvidia.github.io/apex/amp.html" ) , ) parser.add_argument("--n_tpu_cores" , dest="tpu_cores" , type=__A ) parser.add_argument("--max_grad_norm" , dest="gradient_clip_val" , default=1.0 , type=__A , help="Max gradient norm" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_predict" , action="store_true" , help="Whether to run predictions on the test set." ) parser.add_argument( "--gradient_accumulation_steps" , dest="accumulate_grad_batches" , type=__A , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--seed" , type=__A , default=4_2 , help="random seed for initialization" ) parser.add_argument( "--data_dir" , default=str(Path(__A ).parent / "test_run" / "dummy-train-data" ) , type=__A , help="The input data dir. Should contain the training files for the CoNLL-2003 NER task." , ) def _lowerCamelCase( a , a , a=None , a=True , a=[] , a=None , a=None , **a , ): pl.seed_everything(args.seed ) # init model __a = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=__A ) # add custom checkpoints if checkpoint_callback is None: __a = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="checkpoint" , monitor="val_loss" , mode="min" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(__A ) if logging_callback is None: __a = LoggingCallback() __a = {} if args.fpaa: __a = 1_6 if args.gpus > 1: __a = "auto" __a = "ddp" __a = args.accumulate_grad_batches __a = None __a = "auto" __a = pl.Trainer.from_argparse_args( __A , weights_summary=__A , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=__A , val_check_interval=1 , num_sanity_val_steps=2 , **__A , ) if args.do_train: trainer.fit(__A ) else: print("RAG modeling tests with new set functions successfuly executed!" ) return trainer
701
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class snake_case__ : _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : int _snake_case : int _snake_case : float _snake_case : float _snake_case : Tuple[int] def a__ ( self ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def a__ ( self ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def a__ ( self ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def a__ ( self ): __a = torch.arange(self.height * self.width ) __a = torch.stack( [ pixel_indices % self.width, torch.div(lowerCamelCase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def a__ ( self ): __a , *__a = self.shape __a = int(np.prod(lowerCamelCase ) ) __a = self.get_image_coords() __a = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __a = self.get_camera_rays(lowerCamelCase ) __a = rays.view(lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def a__ ( self , lowerCamelCase ): __a , *__a , __a = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __a = coords.view(lowerCamelCase , -1 , 2 ) __a = self.resolution() __a = self.fov() __a = (flat.float() / (res - 1)) * 2 - 1 __a = fracs * torch.tan(fov / 2 ) __a = fracs.view(lowerCamelCase , -1 , 2 ) __a = ( self.z.view(lowerCamelCase , 1 , 3 ) + self.x.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) __a = directions / directions.norm(dim=-1 , keepdim=lowerCamelCase ) __a = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCamelCase , *lowerCamelCase , 2 , 3 ) def a__ ( self , lowerCamelCase , lowerCamelCase ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCamelCase , height=lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowerCamelCase( a ): __a = [] __a = [] __a = [] __a = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __a = np.array([np.sin(a ), np.cos(a ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __a = -z * 4 __a = np.array([np.cos(a ), -np.sin(a ), 0.0] ) __a = np.cross(a , a ) origins.append(a ) xs.append(a ) ys.append(a ) zs.append(a ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(a , axis=0 ) ).float() , x=torch.from_numpy(np.stack(a , axis=0 ) ).float() , y=torch.from_numpy(np.stack(a , axis=0 ) ).float() , z=torch.from_numpy(np.stack(a , axis=0 ) ).float() , width=a , height=a , x_fov=0.7 , y_fov=0.7 , shape=(1, len(a )) , )
67
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class snake_case__ ( unittest.TestCase ): def a__ ( self ): __a = tempfile.mkdtemp() __a = BlipImageProcessor() __a = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-BertModel" ) __a = BlipProcessor(UpperCAmelCase_ , UpperCAmelCase_ ) processor.save_pretrained(self.tmpdirname ) def a__ ( self , **lowerCamelCase ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ).tokenizer def a__ ( self , **lowerCamelCase ): return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ).image_processor def a__ ( self ): shutil.rmtree(self.tmpdirname ) def a__ ( self ): __a = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __a = [Image.fromarray(np.moveaxis(UpperCAmelCase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def a__ ( self ): __a = BlipProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __a = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) __a = self.get_image_processor(do_normalize=UpperCAmelCase_ , padding_value=1.0 ) __a = BlipProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=UpperCAmelCase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase_ ) def a__ ( self ): __a = self.get_image_processor() __a = self.get_tokenizer() __a = BlipProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) __a = self.prepare_image_inputs() __a = image_processor(UpperCAmelCase_ , return_tensors="np" ) __a = processor(images=UpperCAmelCase_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self ): __a = self.get_image_processor() __a = self.get_tokenizer() __a = BlipProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) __a = "lower newer" __a = processor(text=UpperCAmelCase_ ) __a = tokenizer(UpperCAmelCase_ , return_token_type_ids=UpperCAmelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def a__ ( self ): __a = self.get_image_processor() __a = self.get_tokenizer() __a = BlipProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) __a = "lower newer" __a = self.prepare_image_inputs() __a = processor(text=UpperCAmelCase_ , images=UpperCAmelCase_ ) self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] ) # test if it raises when no input is passed with pytest.raises(UpperCAmelCase_ ): processor() def a__ ( self ): __a = self.get_image_processor() __a = self.get_tokenizer() __a = BlipProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) __a = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __a = processor.batch_decode(UpperCAmelCase_ ) __a = tokenizer.batch_decode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def a__ ( self ): __a = self.get_image_processor() __a = self.get_tokenizer() __a = BlipProcessor(tokenizer=UpperCAmelCase_ , image_processor=UpperCAmelCase_ ) __a = "lower newer" __a = self.prepare_image_inputs() __a = processor(text=UpperCAmelCase_ , images=UpperCAmelCase_ ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ["pixel_values", "input_ids", "attention_mask"] )
702
"""simple docstring""" def _lowerCamelCase( a ): return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase( a ): __a = 0 __a = number while duplicate > 0: __a , __a = divmod(a , 1_0 ) fact_sum += factorial(a ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__:Optional[Any] = int(input("""Enter number: """).strip()) print( F'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
67
0
"""simple docstring""" def _lowerCamelCase( a = 5_0_0_0_0_0_0_0 ): __a = set() __a = int((limit - 2_4) ** (1 / 2) ) __a = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , a ) ) ) for primea in primes: __a = primea * primea for primea in primes: __a = primea * primea * primea if square + cube >= limit - 1_6: break for primea in primes: __a = primea * primea * primea * primea __a = square + cube + tetr if total >= limit: break ret.add(a ) return len(a ) if __name__ == "__main__": print(F'''{solution() = }''')
703
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE__:Optional[Any] = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Union[str, Any] = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" class snake_case__ : # Public class to implement a graph def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = row __a = col __a = graph def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __a = [-1, 0, 1, -1, 1, -1, 0, 1] __a = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , UpperCamelCase_ ) def a__ ( self ): # And finally, count all islands. __a = [[False for j in range(self.COL )] for i in range(self.ROW )] __a = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) count += 1 return count
704
"""simple docstring""" import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def _lowerCamelCase( a , a , a ): __a = OmegaConf.load(a ) __a = torch.load(a , map_location="cpu" )["model"] __a = list(state_dict.keys() ) # extract state_dict for VQVAE __a = {} __a = "first_stage_model." for key in keys: if key.startswith(a ): __a = state_dict[key] # extract state_dict for UNetLDM __a = {} __a = "model.diffusion_model." for key in keys: if key.startswith(a ): __a = state_dict[key] __a = config.model.params.first_stage_config.params __a = config.model.params.unet_config.params __a = VQModel(**a ).eval() vqvae.load_state_dict(a ) __a = UNetLDMModel(**a ).eval() unet.load_state_dict(a ) __a = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=a , ) __a = LDMPipeline(a , a , a ) pipeline.save_pretrained(a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
67
0
"""simple docstring""" from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def _lowerCamelCase( a , a , a , a , ): __a = coefficient_matrix.shape __a = constant_matrix.shape if rowsa != colsa: __a = F"Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}" raise ValueError(_A ) if colsa != 1: __a = F"Constant matrix must be nx1 but received {rowsa}x{colsa}" raise ValueError(_A ) if rowsa != rowsa: __a = ( "Coefficient and constant matrices dimensions must be nxn and nx1 but " F"received {rowsa}x{colsa} and {rowsa}x{colsa}" ) raise ValueError(_A ) if len(_A ) != rowsa: __a = ( "Number of initial values must be equal to number of rows in coefficient " F"matrix but received {len(_A )} and {rowsa}" ) raise ValueError(_A ) if iterations <= 0: raise ValueError("Iterations must be at least 1" ) __a = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) __a = table.shape strictly_diagonally_dominant(_A ) # Iterates the whole matrix for given number of times for _ in range(_A ): __a = [] for row in range(_A ): __a = 0 for col in range(_A ): if col == row: __a = table[row][col] elif col == cols - 1: __a = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] __a = (temp + val) / denom new_val.append(_A ) __a = new_val return [float(_A ) for i in new_val] def _lowerCamelCase( a ): __a = table.shape __a = True for i in range(0 , _A ): __a = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError("Coefficient matrix is not strictly diagonally dominant" ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
705
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[Any] = { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json""", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case__ ( snake_case_ ): _snake_case : str = """blenderbot-small""" _snake_case : str = ["""past_key_values"""] _snake_case : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCamelCase=50265 , lowerCamelCase=512 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="gelu" , lowerCamelCase=512 , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , lowerCamelCase=1 , lowerCamelCase=False , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=2 , **lowerCamelCase , ): __a = vocab_size __a = max_position_embeddings __a = d_model __a = encoder_ffn_dim __a = encoder_layers __a = encoder_attention_heads __a = decoder_ffn_dim __a = decoder_layers __a = decoder_attention_heads __a = dropout __a = attention_dropout __a = activation_dropout __a = activation_function __a = init_std __a = encoder_layerdrop __a = decoder_layerdrop __a = use_cache __a = encoder_layers __a = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , is_encoder_decoder=lowerCamelCase , decoder_start_token_id=lowerCamelCase , forced_eos_token_id=lowerCamelCase , **lowerCamelCase , ) class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a = {0: "batch"} __a = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __a = {0: "batch", 1: "decoder_sequence"} __a = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} else: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = super().outputs else: __a = super(lowerCamelCase , self ).outputs if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Generate decoder inputs __a = seq_length if not self.use_past else 1 __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) __a = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} __a = dict(**lowerCamelCase , **lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape __a = common_inputs["decoder_input_ids"].shape[1] __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = decoder_seq_length + 3 __a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __a = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase )] , dim=1 ) __a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __a , __a = self.num_layers __a = min(lowerCamelCase , lowerCamelCase ) __a = max(lowerCamelCase , lowerCamelCase ) - min_num_layers __a = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), ) ) # TODO: test this. __a = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(lowerCamelCase , lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape # Not using the same length for past_key_values __a = seqlen + 2 __a , __a = self.num_layers __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = common_inputs["attention_mask"].dtype __a = torch.cat( [common_inputs["attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) __a = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(lowerCamelCase ) ] return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __a = tokenizer.num_special_tokens_to_add(lowerCamelCase ) __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence __a = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __a = dict(tokenizer(lowerCamelCase , return_tensors=lowerCamelCase ) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): if self.task in ["default", "seq2seq-lm"]: __a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) elif self.task == "causal-lm": __a = self._generate_dummy_inputs_for_causal_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) else: __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if self.task in ["default", "seq2seq-lm"]: __a = super()._flatten_past_key_values_(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __a = super(lowerCamelCase , self )._flatten_past_key_values_( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType SCREAMING_SNAKE_CASE__:Optional[int] = None SCREAMING_SNAKE_CASE__:Union[str, Any] = """<""" if sys.byteorder == """little""" else """>""" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image SCREAMING_SNAKE_CASE__:Optional[int] = [ np.dtype("""|b1"""), np.dtype("""|u1"""), np.dtype("""<u2"""), np.dtype(""">u2"""), np.dtype("""<i2"""), np.dtype(""">i2"""), np.dtype("""<u4"""), np.dtype(""">u4"""), np.dtype("""<i4"""), np.dtype(""">i4"""), np.dtype("""<f4"""), np.dtype(""">f4"""), np.dtype("""<f8"""), np.dtype(""">f8"""), ] @dataclass class snake_case__ : _snake_case : str = True _snake_case : Optional[int] = None # Automatically constructed _snake_case : List[Any] = """PIL.Image.Image""" _snake_case : Tuple = pa.struct({"""bytes""": pa.binary(), """path""": pa.string()} ) _snake_case : Any = field(default="""Image""", init=__lowerCamelCase, repr=__lowerCamelCase ) def __call__( self ): return self.pa_type def a__ ( self , lowerCamelCase ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install \'Pillow\'." ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): __a = np.array(UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): return {"path": value, "bytes": None} elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): return {"path": None, "bytes": value} elif isinstance(UpperCAmelCase_ , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(UpperCAmelCase_ ) elif isinstance(UpperCAmelCase_ , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(UpperCAmelCase_ ) elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( F"An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}." ) def a__ ( self , lowerCamelCase , lowerCamelCase=None ): if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Image(decode=True) instead." ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support decoding images, please install \'Pillow\'." ) if token_per_repo_id is None: __a = {} __a = value['path'], value['bytes'] if bytes_ is None: if path is None: raise ValueError(F"An image should have one of \'path\' or \'bytes\' but both are None in {value}." ) else: if is_local_path(UpperCAmelCase_ ): __a = PIL.Image.open(UpperCAmelCase_ ) else: __a = path.split("::" )[-1] try: __a = string_to_dict(UpperCAmelCase_ , config.HUB_DATASETS_URL )['repo_id'] __a = token_per_repo_id.get(UpperCAmelCase_ ) except ValueError: __a = None with xopen(UpperCAmelCase_ , "rb" , use_auth_token=UpperCAmelCase_ ) as f: __a = BytesIO(f.read() ) __a = PIL.Image.open(bytes_ ) else: __a = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def a__ ( self ): from .features import Value return ( self if self.decode else { "bytes": Value("binary" ), "path": Value("string" ), } ) def a__ ( self , lowerCamelCase ): if pa.types.is_string(storage.type ): __a = pa.array([None] * len(UpperCAmelCase_ ) , type=pa.binary() ) __a = pa.StructArray.from_arrays([bytes_array, storage] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): __a = pa.array([None] * len(UpperCAmelCase_ ) , type=pa.string() ) __a = pa.StructArray.from_arrays([storage, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: __a = storage.field("bytes" ) else: __a = pa.array([None] * len(UpperCAmelCase_ ) , type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: __a = storage.field("path" ) else: __a = pa.array([None] * len(UpperCAmelCase_ ) , type=pa.string() ) __a = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): __a = pa.array( [encode_np_array(np.array(UpperCAmelCase_ ) )["bytes"] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) __a = pa.array([None] * len(UpperCAmelCase_ ) , type=pa.string() ) __a = pa.StructArray.from_arrays( [bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(UpperCAmelCase_ , self.pa_type ) def a__ ( self , lowerCamelCase ): @no_op_if_value_is_null def path_to_bytes(lowerCamelCase ): with xopen(UpperCAmelCase_ , "rb" ) as f: __a = f.read() return bytes_ __a = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) __a = pa.array( [os.path.basename(UpperCAmelCase_ ) if path is not None else None for path in storage.field("path" ).to_pylist()] , type=pa.string() , ) __a = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(UpperCAmelCase_ , self.pa_type ) def _lowerCamelCase( ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install \'Pillow\'." ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() __a = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def _lowerCamelCase( a ): __a = BytesIO() if image.format in list_image_compression_formats(): __a = image.format else: __a = 'PNG' if image.mode in ['1', 'L', 'LA', 'RGB', 'RGBA'] else 'TIFF' image.save(lowerCamelCase_ , format=lowerCamelCase_ ) return buffer.getvalue() def _lowerCamelCase( a ): if hasattr(lowerCamelCase_ , "filename" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(lowerCamelCase_ )} def _lowerCamelCase( a ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install \'Pillow\'." ) __a = array.dtype __a = dtype.byteorder if dtype.byteorder != '=' else _NATIVE_BYTEORDER __a = dtype.kind __a = dtype.itemsize __a = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: __a = np.dtype("|u1" ) if dtype_kind not in ["u", "i"]: raise TypeError( F"Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays." ) if dtype is not dest_dtype: warnings.warn(F"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'" ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: __a = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: __a = dtype_byteorder + dtype_kind + str(lowerCamelCase_ ) __a = np.dtype(lowerCamelCase_ ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F"Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'" ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F"Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}" ) __a = PIL.Image.fromarray(array.astype(lowerCamelCase_ ) ) return {"path": None, "bytes": image_to_bytes(lowerCamelCase_ )} def _lowerCamelCase( a ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install \'Pillow\'." ) if objs: __a = first_non_null_value(lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(lowerCamelCase_ , np.ndarray ): __a = no_op_if_value_is_null(lowerCamelCase_ ) return [obj_to_image_dict_func(lowerCamelCase_ ) for obj in objs] elif isinstance(lowerCamelCase_ , PIL.Image.Image ): __a = no_op_if_value_is_null(lowerCamelCase_ ) return [obj_to_image_dict_func(lowerCamelCase_ ) for obj in objs] else: return objs else: return objs
706
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=99 , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=9 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase=8 , lowerCamelCase=0.1 , lowerCamelCase=0.002 , lowerCamelCase=1 , lowerCamelCase=0 , lowerCamelCase=0 , lowerCamelCase=None , lowerCamelCase=None , ): __a = parent __a = batch_size __a = encoder_seq_length __a = decoder_seq_length # For common tests __a = self.decoder_seq_length __a = is_training __a = use_attention_mask __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = d_ff __a = relative_attention_num_buckets __a = dropout_rate __a = initializer_factor __a = eos_token_id __a = pad_token_id __a = decoder_start_token_id __a = None __a = decoder_layers def a__ ( self ): return TaConfig.from_pretrained("google/umt5-base" ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): if attention_mask is None: __a = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __a = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __a = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=lowerCamelCase ) if decoder_head_mask is None: __a = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) if cross_attn_head_mask is None: __a = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def a__ ( self ): __a = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __a = input_ids.clamp(self.pad_token_id + 1 ) __a = decoder_input_ids.clamp(self.pad_token_id + 1 ) __a = self.get_config() __a = config.num_attention_heads __a = self.prepare_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, input_dict def a__ ( self ): __a , __a = self.prepare_config_and_inputs() return config, inputs_dict def a__ ( self ): return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self ): return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase , attention_mask=lowerCamelCase , decoder_attention_mask=lowerCamelCase , ) __a = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ) __a = result.last_hidden_state __a = result.past_key_values __a = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).get_decoder().to(lowerCamelCase ).eval() # first forward pass __a = model(lowerCamelCase , use_cache=lowerCamelCase ) __a = model(lowerCamelCase ) __a = model(lowerCamelCase , use_cache=lowerCamelCase ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) + 1 ) __a , __a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = model(lowerCamelCase )["last_hidden_state"] __a = model(lowerCamelCase , past_key_values=lowerCamelCase )["last_hidden_state"] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -1, random_slice_idx].detach() __a = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).to(lowerCamelCase ).half().eval() __a = model(**lowerCamelCase )["last_hidden_state"] self.parent.assertFalse(torch.isnan(lowerCamelCase ).any().item() ) @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _snake_case : int = (UMTaForConditionalGeneration,) if is_torch_available() else () _snake_case : Optional[int] = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _snake_case : List[Any] = True _snake_case : Union[str, Any] = False _snake_case : Union[str, Any] = False _snake_case : Tuple = True _snake_case : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests _snake_case : Optional[Any] = [0.8, 0.9] def a__ ( self ): __a = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() __a = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=lowerCamelCase , opset_version=9 , input_names=["input_ids", "decoder_input_ids"] , ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase ) def a__ ( self ): __a = ["encoder_attentions", "decoder_attentions", "cross_attentions"] __a = self.model_tester.prepare_config_and_inputs() __a = config_and_inputs[0] __a = UMTaForConditionalGeneration(lowerCamelCase ).eval() model.to(lowerCamelCase ) __a = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=lowerCamelCase ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), } for attn_name, (name, mask) in zip(lowerCamelCase , head_masking.items() ): __a = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": __a = torch.ones( config.num_decoder_layers , config.num_heads , device=lowerCamelCase ) __a = model.generate( config_and_inputs[1]["input_ids"] , num_beams=1 , max_length=3 , output_attentions=lowerCamelCase , return_dict_in_generate=lowerCamelCase , **lowerCamelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step __a = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def a__ ( self ): pass @require_torch @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def a__ ( self ): __a = UMTaForConditionalGeneration.from_pretrained("google/umt5-small" , return_dict=lowerCamelCase ).to(lowerCamelCase ) __a = AutoTokenizer.from_pretrained("google/umt5-small" , use_fast=lowerCamelCase , legacy=lowerCamelCase ) __a = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] __a = tokenizer(lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ).input_ids # fmt: off __a = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase , lowerCamelCase ) __a = model.generate(input_ids.to(lowerCamelCase ) ) __a = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] __a = tokenizer.batch_decode(lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" def _lowerCamelCase( a , a ): return int((input_a, input_a).count(1 ) != 0 ) def _lowerCamelCase( ): assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
707
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase( a , a , a ): # Initialise PyTorch model __a = MobileBertConfig.from_json_file(a ) print(F"Building PyTorch model from configuration: {config}" ) __a = MobileBertForPreTraining(a ) # Load weights from tf checkpoint __a = load_tf_weights_in_mobilebert(a , a , a ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) SCREAMING_SNAKE_CASE__:List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
67
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE__:List[Any] = { """configuration_m2m_100""": ["""M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP""", """M2M100Config""", """M2M100OnnxConfig"""], """tokenization_m2m_100""": ["""M2M100Tokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:int = [ """M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST""", """M2M100ForConditionalGeneration""", """M2M100Model""", """M2M100PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
708
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__ ( snake_case_ ): def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __a = input_file.read() __a = regexp.search(lowerCamelCase ) return match def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __a = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __a = regexp.finditer(lowerCamelCase ) __a = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCamelCase ) ): raise AssertionError(F"open(...) must use utf-8 encoding in {dataset}" ) def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCamelCase ) ): raise AssertionError(F"print statement found in {dataset}. Use datasets.logger/logging instead." )
67
0
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : Dict = DiTPipeline _snake_case : Dict = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS _snake_case : Optional[Any] = PipelineTesterMixin.required_optional_params - { """latents""", """num_images_per_prompt""", """callback""", """callback_steps""", } _snake_case : List[Any] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS _snake_case : Dict = False def a__ ( self ): torch.manual_seed(0 ) __a = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=_lowerCAmelCase , activation_fn="gelu-approximate" , num_embeds_ada_norm=1000 , norm_type="ada_norm_zero" , norm_elementwise_affine=_lowerCAmelCase , ) __a = AutoencoderKL() __a = DDIMScheduler() __a = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def a__ ( self , lowerCamelCase , lowerCamelCase=0 ): if str(_lowerCAmelCase ).startswith("mps" ): __a = torch.manual_seed(_lowerCAmelCase ) else: __a = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) __a = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def a__ ( self ): __a = "cpu" __a = self.get_dummy_components() __a = self.pipeline_class(**_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) __a = self.get_dummy_inputs(_lowerCAmelCase ) __a = pipe(**_lowerCAmelCase ).images __a = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) __a = np.array([0.2946, 0.6601, 0.4329, 0.3296, 0.4144, 0.5319, 0.7273, 0.5013, 0.4457] ) __a = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_lowerCAmelCase , 1E-3 ) def a__ ( self ): self._test_inference_batch_single_identical(relax_max_difference=_lowerCAmelCase , expected_max_diff=1E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def a__ ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @require_torch_gpu @slow class snake_case__ ( unittest.TestCase ): def a__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = torch.manual_seed(0 ) __a = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) __a = ["vase", "umbrella", "white shark", "white wolf"] __a = pipe.get_label_ids(_lowerCAmelCase ) __a = pipe(_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=40 , output_type="np" ).images for word, image in zip(_lowerCAmelCase , _lowerCAmelCase ): __a = load_numpy( F"https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy" ) assert np.abs((expected_image - image).max() ) < 1E-2 def a__ ( self ): __a = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) __a = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) __a = ["vase", "umbrella"] __a = pipe.get_label_ids(_lowerCAmelCase ) __a = torch.manual_seed(0 ) __a = pipe(_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=25 , output_type="np" ).images for word, image in zip(_lowerCAmelCase , _lowerCAmelCase ): __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" F"/dit/{word}_512.npy" ) assert np.abs((expected_image - image).max() ) < 1E-1
709
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
67
0
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() SCREAMING_SNAKE_CASE__:List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Union[str, Any] = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "ctc_proj", "mask_emb": "masked_spec_embed", } SCREAMING_SNAKE_CASE__:Optional[int] = [ "ctc_proj", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def _lowerCamelCase( a , a , a , a , a , a ) -> Optional[int]: for attribute in key.split("." ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models __a = "lm_head" __a = getattr(lowerCamelCase__ , lowerCamelCase__ ) if weight_type is not None: __a = getattr(lowerCamelCase__ , lowerCamelCase__ ).shape else: __a = hf_pointer.shape assert hf_shape == value.shape, ( F"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" F" {value.shape} for {full_name}" ) if weight_type == "weight": __a = value elif weight_type == "weight_g": __a = value elif weight_type == "weight_v": __a = value elif weight_type == "bias": __a = value else: __a = value logger.info(F"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def _lowerCamelCase( a , a , a ) -> Tuple: __a = [] __a = fairseq_model.state_dict() __a = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): __a = False if "conv_layers" in name: load_conv_layer( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , hf_model.config.feat_extract_norm == "group" , ) __a = True else: for key, mapped_key in MAPPING.items(): __a = "unispeech." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: __a = True if "*" in mapped_key: __a = name.split(lowerCamelCase__ )[0].split("." )[-2] __a = mapped_key.replace("*" , lowerCamelCase__ ) if "weight_g" in name: __a = "weight_g" elif "weight_v" in name: __a = "weight_v" elif "bias" in name: __a = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj __a = "weight" else: __a = None set_recursively(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) continue if not is_used: unused_weights.append(lowerCamelCase__ ) logger.warning(F"Unused weights: {unused_weights}" ) def _lowerCamelCase( a , a , a , a , a ) -> Any: __a = full_name.split("conv_layers." )[-1] __a = name.split("." ) __a = int(items[0] ) __a = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) __a = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) __a = value logger.info(F"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was" " found." ) __a = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"{full_name} has size {value.shape}, but" F" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found." ) __a = value logger.info(F"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(lowerCamelCase__ ) @torch.no_grad() def _lowerCamelCase( a , a , a=None , a=None , a=True ) -> Union[str, Any]: if config_path is not None: __a = UniSpeechConfig.from_pretrained(lowerCamelCase__ ) else: __a = UniSpeechConfig() if is_finetuned: if dict_path: __a = Dictionary.load_from_json(lowerCamelCase__ ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __a = target_dict.pad_index __a = target_dict.bos_index __a = target_dict.eos_index __a = len(target_dict.symbols ) __a = os.path.join(lowerCamelCase__ , "vocab.json" ) if not os.path.isdir(lowerCamelCase__ ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(lowerCamelCase__ ) ) return os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) __a = target_dict.indices # fairseq has the <pad> and <s> switched __a = 4_2 __a = 4_3 with open(lowerCamelCase__ , "w" , encoding="utf-8" ) as vocab_handle: json.dump(lowerCamelCase__ , lowerCamelCase__ ) __a = WavaVecaPhonemeCTCTokenizer( lowerCamelCase__ , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=lowerCamelCase__ , ) __a = True if config.feat_extract_norm == "layer" else False __a = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0 , do_normalize=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , ) __a = WavaVecaProcessor(feature_extractor=lowerCamelCase__ , tokenizer=lowerCamelCase__ ) processor.save_pretrained(lowerCamelCase__ ) __a = UniSpeechForCTC(lowerCamelCase__ ) else: __a = UniSpeechForPreTraining(lowerCamelCase__ ) if is_finetuned: __a = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] ), "w2v_path": checkpoint_path} ) else: __a = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __a = model[0].eval() recursively_load_weights(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) hf_unispeech.save_pretrained(lowerCamelCase__ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) SCREAMING_SNAKE_CASE__:Optional[int] = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
710
"""simple docstring""" import heapq import sys import numpy as np SCREAMING_SNAKE_CASE__:Optional[int] = tuple[int, int] class snake_case__ : def __init__( self ): __a = [] __a = set() def a__ ( self ): if not self.empty(): return self.elements[0][0] else: return float("inf" ) def a__ ( self ): return len(self.elements ) == 0 def a__ ( self , lowerCamelCase , lowerCamelCase ): if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(lowerCamelCase ) else: # update # print("update", item) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def a__ ( self , lowerCamelCase ): if item in self.set: self.set.remove(lowerCamelCase ) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def a__ ( self ): return self.elements[0][1] def a__ ( self ): ((__a) , (__a)) = heapq.heappop(self.elements ) self.set.remove(lowerCamelCase ) return (priority, item) def _lowerCamelCase( a , a ): # euclidean distance __a = np.array(a ) __a = np.array(a ) return np.linalg.norm(a - b ) def _lowerCamelCase( a , a ): # integer division by time variable return consistent_heuristic(a , a ) // t def _lowerCamelCase( a , a ): # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _lowerCamelCase( a , a , a , a ): __a = g_function[start] + Wa * heuristics[i](a , a ) return ans def _lowerCamelCase( a , a , a ): __a = np.chararray((n, n) ) for i in range(a ): for j in range(a ): __a = "*" for i in range(a ): for j in range(a ): if (j, (n - 1) - i) in blocks: __a = "#" __a = "-" __a = back_pointer[goal] while x != start: ((__a) , (__a)) = x # print(x) __a = "-" __a = back_pointer[x] __a = "-" for i in range(a ): for j in range(a ): if (i, j) == (0, n - 1): print(grid[i][j] , end=" " ) print("<-- End position" , end=" " ) else: print(grid[i][j] , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) print("PATH TAKEN BY THE ALGORITHM IS:-" ) __a = back_pointer[goal] while x != start: print(a , end=" " ) __a = back_pointer[x] print(a ) sys.exit() def _lowerCamelCase( a ): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _lowerCamelCase( a , a , a , a , a , a , a , a , ): for itera in range(a ): open_list[itera].remove_element(a ) # print("s", s) # print("j", j) ((__a) , (__a)) = s __a = (x - 1, y) __a = (x + 1, y) __a = (x, y + 1) __a = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(a ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(a ) __a = -1 __a = float("inf" ) if valid(a ) and g_function[neighbours] > g_function[s] + 1: __a = g_function[s] + 1 __a = s if neighbours not in close_list_anchor: open_list[0].put(a , key(a , 0 , a , a ) ) if neighbours not in close_list_inad: for var in range(1 , a ): if key(a , a , a , a ) <= Wa * key( a , 0 , a , a ): open_list[j].put( a , key(a , a , a , a ) ) def _lowerCamelCase( ): __a = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(1_5 , 2_0 ): some_list.append((x, 1_7) ) for x in range(1_0 , 1_9 ): for y in range(1 , 1_5 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(1_2 , 1_9 ): some_list.append((x, y) ) for x in range(3 , 1_3 ): for y in range(1_6 , 1_9 ): some_list.append((x, y) ) return some_list SCREAMING_SNAKE_CASE__:Any = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} SCREAMING_SNAKE_CASE__:str = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] SCREAMING_SNAKE_CASE__:int = make_common_ground() SCREAMING_SNAKE_CASE__:List[str] = blocks_blk # hyper parameters SCREAMING_SNAKE_CASE__:str = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 20 SCREAMING_SNAKE_CASE__:Dict = 3 # one consistent and two other inconsistent # start and end destination SCREAMING_SNAKE_CASE__:Dict = (0, 0) SCREAMING_SNAKE_CASE__:Optional[Any] = (n - 1, n - 1) SCREAMING_SNAKE_CASE__:List[str] = 1 def _lowerCamelCase( a , a , a ): __a = {start: 0, goal: float("inf" )} __a = {start: -1, goal: -1} __a = [] __a = set() for i in range(a ): open_list.append(PriorityQueue() ) open_list[i].put(a , key(a , a , a , a ) ) __a = [] __a = [] while open_list[0].minkey() < float("inf" ): for i in range(1 , a ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a , __a = open_list[i].top_show() visited.add(a ) expand_state( a , a , a , a , a , a , a , a , ) close_list_inad.append(a ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a = open_list[0].top_show() visited.add(a ) expand_state( a , 0 , a , a , a , a , a , a , ) close_list_anchor.append(a ) print("No path found to goal" ) print() for i in range(n - 1 , -1 , -1 ): for j in range(a ): if (j, i) in blocks: print("#" , end=" " ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" " ) else: print("-" , end=" " ) else: print("*" , end=" " ) if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
67
0
"""simple docstring""" from math import loga def _lowerCamelCase( a ): if a < 0: raise ValueError("Input value must be a positive integer" ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("Input value must be a 'int' type" ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
711
"""simple docstring""" SCREAMING_SNAKE_CASE__:Any = """Alexander Joslin""" import operator as op from .stack import Stack def _lowerCamelCase( a ): __a = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} __a = Stack() __a = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a ) ) elif i in operators: # RULE 2 operator_stack.push(a ) elif i == ")": # RULE 4 __a = operator_stack.peek() operator_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operators[opr](a , a ) operand_stack.push(a ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
67
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__:str = logging.get_logger(__name__) def _lowerCamelCase( a ): __a = DPTConfig(embedding_type="hybrid" ) if "large" in checkpoint_url: __a = 1_0_2_4 __a = 4_0_9_6 __a = 2_4 __a = 1_6 __a = [5, 1_1, 1_7, 2_3] __a = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] __a = (1, 3_8_4, 3_8_4) if "nyu" or "midas" in checkpoint_url: __a = 7_6_8 __a = [1, 1, 1, 0.5] __a = [2_5_6, 5_1_2, 7_6_8, 7_6_8] __a = 1_5_0 __a = 1_6 __a = (1, 3_8_4, 3_8_4) __a = False __a = """project""" if "ade" in checkpoint_url: __a = True __a = 7_6_8 __a = [1, 1, 1, 0.5] __a = 1_5_0 __a = 1_6 __a = """huggingface/label-files""" __a = """ade20k-id2label.json""" __a = json.load(open(cached_download(hf_hub_url(__lowerCAmelCase , __lowerCAmelCase , repo_type="dataset" ) ) , "r" ) ) __a = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} __a = idalabel __a = {v: k for k, v in idalabel.items()} __a = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCamelCase( a ): __a = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def _lowerCamelCase( a ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): __a = name.replace("pretrained.model" , "dpt.encoder" ) if "pretrained.model" in name: __a = name.replace("pretrained.model" , "dpt.embeddings" ) if "patch_embed" in name: __a = name.replace("patch_embed" , "" ) if "pos_embed" in name: __a = name.replace("pos_embed" , "position_embeddings" ) if "attn.proj" in name: __a = name.replace("attn.proj" , "attention.output.dense" ) if "proj" in name and "project" not in name: __a = name.replace("proj" , "projection" ) if "blocks" in name: __a = name.replace("blocks" , "layer" ) if "mlp.fc1" in name: __a = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __a = name.replace("mlp.fc2" , "output.dense" ) if "norm1" in name and "backbone" not in name: __a = name.replace("norm1" , "layernorm_before" ) if "norm2" in name and "backbone" not in name: __a = name.replace("norm2" , "layernorm_after" ) if "scratch.output_conv" in name: __a = name.replace("scratch.output_conv" , "head" ) if "scratch" in name: __a = name.replace("scratch" , "neck" ) if "layer1_rn" in name: __a = name.replace("layer1_rn" , "convs.0" ) if "layer2_rn" in name: __a = name.replace("layer2_rn" , "convs.1" ) if "layer3_rn" in name: __a = name.replace("layer3_rn" , "convs.2" ) if "layer4_rn" in name: __a = name.replace("layer4_rn" , "convs.3" ) if "refinenet" in name: __a = int(name[len("neck.refinenet" ) : len("neck.refinenet" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 __a = name.replace(F"refinenet{layer_idx}" , F"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: __a = name.replace("out_conv" , "projection" ) if "resConfUnit1" in name: __a = name.replace("resConfUnit1" , "residual_layer1" ) if "resConfUnit2" in name: __a = name.replace("resConfUnit2" , "residual_layer2" ) if "conv1" in name: __a = name.replace("conv1" , "convolution1" ) if "conv2" in name: __a = name.replace("conv2" , "convolution2" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: __a = name.replace("pretrained.act_postprocess1.0.project.0" , "neck.reassemble_stage.readout_projects.0.0" ) if "pretrained.act_postprocess2.0.project.0" in name: __a = name.replace("pretrained.act_postprocess2.0.project.0" , "neck.reassemble_stage.readout_projects.1.0" ) if "pretrained.act_postprocess3.0.project.0" in name: __a = name.replace("pretrained.act_postprocess3.0.project.0" , "neck.reassemble_stage.readout_projects.2.0" ) if "pretrained.act_postprocess4.0.project.0" in name: __a = name.replace("pretrained.act_postprocess4.0.project.0" , "neck.reassemble_stage.readout_projects.3.0" ) # resize blocks if "pretrained.act_postprocess1.3" in name: __a = name.replace("pretrained.act_postprocess1.3" , "neck.reassemble_stage.layers.0.projection" ) if "pretrained.act_postprocess1.4" in name: __a = name.replace("pretrained.act_postprocess1.4" , "neck.reassemble_stage.layers.0.resize" ) if "pretrained.act_postprocess2.3" in name: __a = name.replace("pretrained.act_postprocess2.3" , "neck.reassemble_stage.layers.1.projection" ) if "pretrained.act_postprocess2.4" in name: __a = name.replace("pretrained.act_postprocess2.4" , "neck.reassemble_stage.layers.1.resize" ) if "pretrained.act_postprocess3.3" in name: __a = name.replace("pretrained.act_postprocess3.3" , "neck.reassemble_stage.layers.2.projection" ) if "pretrained.act_postprocess4.3" in name: __a = name.replace("pretrained.act_postprocess4.3" , "neck.reassemble_stage.layers.3.projection" ) if "pretrained.act_postprocess4.4" in name: __a = name.replace("pretrained.act_postprocess4.4" , "neck.reassemble_stage.layers.3.resize" ) if "pretrained" in name: __a = name.replace("pretrained" , "dpt" ) if "bn" in name: __a = name.replace("bn" , "batch_norm" ) if "head" in name: __a = name.replace("head" , "head.head" ) if "encoder.norm" in name: __a = name.replace("encoder.norm" , "layernorm" ) if "auxlayer" in name: __a = name.replace("auxlayer" , "auxiliary_head.head" ) if "backbone" in name: __a = name.replace("backbone" , "backbone.bit.encoder" ) if ".." in name: __a = name.replace(".." , "." ) if "stem.conv" in name: __a = name.replace("stem.conv" , "bit.embedder.convolution" ) if "blocks" in name: __a = name.replace("blocks" , "layers" ) if "convolution" in name and "backbone" in name: __a = name.replace("convolution" , "conv" ) if "layer" in name and "backbone" in name: __a = name.replace("layer" , "layers" ) if "backbone.bit.encoder.bit" in name: __a = name.replace("backbone.bit.encoder.bit" , "backbone.bit" ) if "embedder.conv" in name: __a = name.replace("embedder.conv" , "embedder.convolution" ) if "backbone.bit.encoder.stem.norm" in name: __a = name.replace("backbone.bit.encoder.stem.norm" , "backbone.bit.embedder.norm" ) return name def _lowerCamelCase( a , a ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __a = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.weight" ) __a = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __a = in_proj_weight[: config.hidden_size, :] __a = in_proj_bias[: config.hidden_size] __a = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __a = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __a = in_proj_weight[ -config.hidden_size :, : ] __a = in_proj_bias[-config.hidden_size :] def _lowerCamelCase( ): __a = """http://images.cocodataset.org/val2017/000000039769.jpg""" __a = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def _lowerCamelCase( a , a , a , a , a ): __a = get_dpt_config(__lowerCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") __a = torch.load(__lowerCAmelCase , map_location="cpu" ) # remove certain keys remove_ignore_keys_(__lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): __a = state_dict.pop(__lowerCAmelCase ) __a = val # read in qkv matrices read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model __a = DPTForSemanticSegmentation(__lowerCAmelCase ) if """ade""" in checkpoint_url else DPTForDepthEstimation(__lowerCAmelCase ) model.load_state_dict(__lowerCAmelCase ) model.eval() # Check outputs on an image __a = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 __a = DPTImageProcessor(size=__lowerCAmelCase ) __a = prepare_img() __a = image_processor(__lowerCAmelCase , return_tensors="pt" ) # forward pass __a = model(**__lowerCAmelCase ).logits if """ade""" in checkpoint_url else model(**__lowerCAmelCase ).predicted_depth if show_prediction: __a = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode="bicubic" , align_corners=__lowerCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 2_5_5 ).show() if pytorch_dump_folder_path is not None: Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(__lowerCAmelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__lowerCAmelCase ) if push_to_hub: model.push_to_hub("ybelkada/dpt-hybrid-midas" ) image_processor.push_to_hub("ybelkada/dpt-hybrid-midas" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:str = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint_url""", default="""https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt""", type=str, help="""URL of the original DPT checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) parser.add_argument( """--model_name""", default="""dpt-large""", type=str, help="""Name of the model, in case you're pushing to the hub.""", ) parser.add_argument( """--show_prediction""", action="""store_true""", ) SCREAMING_SNAKE_CASE__:int = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
712
"""simple docstring""" from math import pi def _lowerCamelCase( a , a ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
67
0
import argparse import math import os import torch from neural_compressor.utils.pytorch import load from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, StableDiffusionPipeline, UNetaDConditionModel def _lowerCamelCase( ): __a = argparse.ArgumentParser() parser.add_argument( "-m" , "--pretrained_model_name_or_path" , type=__UpperCAmelCase , default=__UpperCAmelCase , required=__UpperCAmelCase , help="Path to pretrained model or model identifier from huggingface.co/models." , ) parser.add_argument( "-c" , "--caption" , type=__UpperCAmelCase , default="robotic cat with wings" , help="Text used to generate images." , ) parser.add_argument( "-n" , "--images_num" , type=__UpperCAmelCase , default=4 , help="How much images to generate." , ) parser.add_argument( "-s" , "--seed" , type=__UpperCAmelCase , default=4_2 , help="Seed for random process." , ) parser.add_argument( "-ci" , "--cuda_id" , type=__UpperCAmelCase , default=0 , help="cuda_id." , ) __a = parser.parse_args() return args def _lowerCamelCase( a , a , a ): if not len(__UpperCAmelCase ) == rows * cols: raise ValueError("The specified number of rows and columns are not correct." ) __a , __a = imgs[0].size __a = Image.new("RGB" , size=(cols * w, rows * h) ) __a , __a = grid.size for i, img in enumerate(__UpperCAmelCase ): grid.paste(__UpperCAmelCase , box=(i % cols * w, i // cols * h) ) return grid def _lowerCamelCase( a , a="robotic cat with wings" , a=7.5 , a=5_0 , a=1 , a=4_2 , ): __a = torch.Generator(pipeline.device ).manual_seed(__UpperCAmelCase ) __a = pipeline( __UpperCAmelCase , guidance_scale=__UpperCAmelCase , num_inference_steps=__UpperCAmelCase , generator=__UpperCAmelCase , num_images_per_prompt=__UpperCAmelCase , ).images __a = int(math.sqrt(__UpperCAmelCase ) ) __a = image_grid(__UpperCAmelCase , rows=_rows , cols=num_images_per_prompt // _rows ) return grid, images SCREAMING_SNAKE_CASE__:str = parse_args() # Load models and create wrapper for stable diffusion SCREAMING_SNAKE_CASE__:Any = CLIPTokenizer.from_pretrained(args.pretrained_model_name_or_path, subfolder="""tokenizer""") SCREAMING_SNAKE_CASE__:Optional[int] = CLIPTextModel.from_pretrained(args.pretrained_model_name_or_path, subfolder="""text_encoder""") SCREAMING_SNAKE_CASE__:Optional[Any] = AutoencoderKL.from_pretrained(args.pretrained_model_name_or_path, subfolder="""vae""") SCREAMING_SNAKE_CASE__:Optional[int] = UNetaDConditionModel.from_pretrained(args.pretrained_model_name_or_path, subfolder="""unet""") SCREAMING_SNAKE_CASE__:Tuple = StableDiffusionPipeline.from_pretrained( args.pretrained_model_name_or_path, text_encoder=text_encoder, vae=vae, unet=unet, tokenizer=tokenizer ) SCREAMING_SNAKE_CASE__:Optional[int] = lambda images, clip_input: (images, False) if os.path.exists(os.path.join(args.pretrained_model_name_or_path, """best_model.pt""")): SCREAMING_SNAKE_CASE__:Tuple = load(args.pretrained_model_name_or_path, model=unet) unet.eval() setattr(pipeline, """unet""", unet) else: SCREAMING_SNAKE_CASE__:Optional[Any] = unet.to(torch.device("""cuda""", args.cuda_id)) SCREAMING_SNAKE_CASE__:str = pipeline.to(unet.device) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__:List[str] = generate_images(pipeline, prompt=args.caption, num_images_per_prompt=args.images_num, seed=args.seed) grid.save(os.path.join(args.pretrained_model_name_or_path, """{}.png""".format("""_""".join(args.caption.split())))) SCREAMING_SNAKE_CASE__:List[Any] = os.path.join(args.pretrained_model_name_or_path, """_""".join(args.caption.split())) os.makedirs(dirname, exist_ok=True) for idx, image in enumerate(images): image.save(os.path.join(dirname, """{}.png""".format(idx + 1)))
713
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): _snake_case : Dict = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = IMAGENET_DEFAULT_MEAN , lowerCamelCase = IMAGENET_DEFAULT_STD , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __a = image_std if image_std is not None else IMAGENET_DEFAULT_STD def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __a = int((256 / 224) * size["shortest_edge"] ) __a = get_resize_output_image_size(lowerCamelCase , size=lowerCamelCase , default_to_square=lowerCamelCase ) __a = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}" ) return resize( lowerCamelCase , size=(size_dict["height"], size_dict["width"]) , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"Size dict must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(lowerCamelCase , lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(lowerCamelCase , lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
"""simple docstring""" import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL SCREAMING_SNAKE_CASE__:Optional[Any] = version.parse(version.parse(torch.__version__).base_version) < version.parse("""1.11""") def _lowerCamelCase( a , a , a , a , a , a , a , a=False , ): output_path.parent.mkdir(parents=lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , use_external_data_format=lowerCAmelCase__ , enable_onnx_checker=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) else: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) @torch.no_grad() def _lowerCamelCase( a , a , a , a = False ): __a = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): __a = "cuda" elif fpaa and not torch.cuda.is_available(): raise ValueError("`float16` model export is only supported on GPUs with CUDA" ) else: __a = "cpu" __a = Path(lowerCAmelCase__ ) # VAE DECODER __a = AutoencoderKL.from_pretrained(model_path + "/vae" ) __a = vae_decoder.config.latent_channels # forward only through the decoder part __a = vae_decoder.decode onnx_export( lowerCAmelCase__ , model_args=( torch.randn(1 , lowerCAmelCase__ , 2_5 , 2_5 ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=output_path / "vae_decoder" / "model.onnx" , ordered_input_names=["latent_sample", "return_dict"] , output_names=["sample"] , dynamic_axes={ "latent_sample": {0: "batch", 1: "channels", 2: "height", 3: "width"}, } , opset=lowerCAmelCase__ , ) del vae_decoder if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Any = argparse.ArgumentParser() parser.add_argument( """--model_path""", type=str, required=True, help="""Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).""", ) parser.add_argument("""--output_path""", type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--opset""", default=14, type=int, help="""The version of the ONNX operator set to use.""", ) parser.add_argument("""--fp16""", action="""store_true""", default=False, help="""Export the models in `float16` mode""") SCREAMING_SNAKE_CASE__:Optional[int] = parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print("""SD: Done: ONNX""")
714
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=30 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=None , lowerCamelCase=2 , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def a__ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, pixel_values, labels def a__ ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTForMaskedImageModeling(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __a = 1 __a = ViTForMaskedImageModeling(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.type_sequence_label_size __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a = 1 __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, unittest.TestCase ): _snake_case : Any = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) _snake_case : List[Any] = ( {"""feature-extraction""": ViTModel, """image-classification""": ViTForImageClassification} if is_torch_available() else {} ) _snake_case : int = True _snake_case : int = False _snake_case : str = False _snake_case : Optional[Any] = False def a__ ( self ): __a = ViTModelTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def a__ ( self ): pass def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def a__ ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def _lowerCamelCase( ): __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def a__ ( self ): return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def a__ ( self ): __a = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(lowerCamelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(**lowerCamelCase ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __a = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def a__ ( self ): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. __a = ViTModel.from_pretrained("facebook/dino-vits8" ).to(lowerCamelCase ) __a = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(lowerCamelCase , interpolate_pos_encoding=lowerCamelCase ) # verify the logits __a = torch.Size((1, 3601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , lowerCamelCase ) __a = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def a__ ( self ): __a = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __a = model(lowerCamelCase )
67
0
"""simple docstring""" def _lowerCamelCase( a ): __a = 1 for i in range(1 , num + 1 ): fact *= i return fact def _lowerCamelCase( a ): __a = 0 while number > 0: __a = number % 1_0 sum_of_digits += last_digit __a = number // 1_0 # Removing the last_digit from the given number return sum_of_digits def _lowerCamelCase( a = 1_0_0 ): __a = factorial(__A ) __a = split_and_add(__A ) return result if __name__ == "__main__": print(solution(int(input("""Enter the Number: """).strip())))
715
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=True , lowerCamelCase=1 / 255 , lowerCamelCase=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __a = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __a = parent __a = batch_size __a = num_channels __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std __a = do_rescale __a = rescale_factor __a = do_pad def a__ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def a__ ( self , lowerCamelCase , lowerCamelCase=False ): if not batched: __a = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __a , __a = image.size else: __a , __a = image.shape[1], image.shape[2] if w < h: __a = int(self.size["shortest_edge"] * h / w ) __a = self.size["shortest_edge"] elif w > h: __a = self.size["shortest_edge"] __a = int(self.size["shortest_edge"] * w / h ) else: __a = self.size["shortest_edge"] __a = self.size["shortest_edge"] else: __a = [] for image in image_inputs: __a , __a = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : List[Any] = DetaImageProcessor if is_vision_available() else None def a__ ( self ): __a = DetaImageProcessingTester(self ) @property def a__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_rescale" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_pad" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def a__ ( self ): __a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def a__ ( self ): pass def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self ): # prepare image and target __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"image_id": 39769, "annotations": target} # encode them __a = DetaImageProcessor() __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def a__ ( self ): # prepare image, target and masks_path __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __a = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __a = DetaImageProcessor(format="coco_panoptic" ) __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __a = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
67
0
"""simple docstring""" def _lowerCamelCase( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(_lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(F'''{solution() = }''')
716
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE__:Dict = logging.getLogger() def _lowerCamelCase( ): __a = argparse.ArgumentParser() parser.add_argument("-f" ) __a = parser.parse_args() return args.f class snake_case__ ( snake_case_ ): def a__ ( self ): __a = logging.StreamHandler(sys.stdout ) logger.addHandler(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowerCamelCase , "argv" , lowerCamelCase ): __a = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self ): __a = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase )
67
0
from __future__ import annotations def _lowerCamelCase( a , a ): __a = sorted(numsa + numsa ) __a = divmod(len(_lowerCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE__:Tuple = [float(x) for x in input("""Enter the elements of first array: """).split()] SCREAMING_SNAKE_CASE__:Union[str, Any] = [float(x) for x in input("""Enter the elements of second array: """).split()] print(F'''The median of two arrays is: {median_of_two_arrays(array_a, array_a)}''')
717
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __a = image_std if image_std is not None else OPENAI_CLIP_STD __a = do_convert_rgb def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) __a = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , param_name="size" , default_to_square=lowerCamelCase ) __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" , default_to_square=lowerCamelCase ) __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __a = [convert_to_rgb(lowerCamelCase ) for image in images] # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=128 , lowerCamelCase=32 , lowerCamelCase=16 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=3 , lowerCamelCase=4 , lowerCamelCase=None , ): __a = parent __a = batch_size __a = seq_length __a = is_training __a = use_input_mask __a = use_token_type_ids __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = type_vocab_size __a = type_sequence_label_size __a = initializer_range __a = num_labels __a = num_choices __a = scope def a__ ( self ): __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_input_mask: __a = random_attention_mask([self.batch_size, self.seq_length] ) __a = None if self.use_token_type_ids: __a = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __a = None __a = None __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __a = ids_tensor([self.batch_size] , self.num_choices ) __a = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a__ ( self ): return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase_ , initializer_range=self.initializer_range , ) def a__ ( self ): ( __a ) = self.prepare_config_and_inputs() __a = True __a = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = NezhaModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ ) __a = model(lowerCamelCase_ , token_type_ids=lowerCamelCase_ ) __a = model(lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = True __a = NezhaModel(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model( lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , encoder_hidden_states=lowerCamelCase_ , encoder_attention_mask=lowerCamelCase_ , ) __a = model( lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , encoder_hidden_states=lowerCamelCase_ , ) __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = NezhaForMaskedLM(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = NezhaForNextSentencePrediction(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model( lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , labels=lowerCamelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = NezhaForPreTraining(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model( lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , labels=lowerCamelCase_ , next_sentence_label=lowerCamelCase_ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = NezhaForQuestionAnswering(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model( lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , start_positions=lowerCamelCase_ , end_positions=lowerCamelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.num_labels __a = NezhaForSequenceClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.num_labels __a = NezhaForTokenClassification(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.num_choices __a = NezhaForMultipleChoice(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() __a = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __a = model( lowerCamelCase_ , attention_mask=lowerCamelCase_ , token_type_ids=lowerCamelCase_ , labels=lowerCamelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def a__ ( self ): __a = self.prepare_config_and_inputs() ( __a ) = config_and_inputs __a = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class snake_case__ ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase, unittest.TestCase ): _snake_case : int = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) _snake_case : List[str] = ( { """feature-extraction""": NezhaModel, """fill-mask""": NezhaForMaskedLM, """question-answering""": NezhaForQuestionAnswering, """text-classification""": NezhaForSequenceClassification, """token-classification""": NezhaForTokenClassification, """zero-shot""": NezhaForSequenceClassification, } if is_torch_available() else {} ) _snake_case : Union[str, Any] = True def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=False ): __a = super()._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ , return_labels=lowerCamelCase_ ) if return_labels: if model_class in get_values(lowerCamelCase_ ): __a = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowerCamelCase_ ) __a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase_ ) return inputs_dict def a__ ( self ): __a = NezhaModelTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase_ , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCamelCase_ ) def a__ ( self ): ( __a ) = self.model_tester.prepare_config_and_inputs_for_decoder() __a = None self.model_tester.create_and_check_model_as_decoder( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase_ ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase_ ) @slow def a__ ( self ): for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = NezhaModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) @slow @require_torch_gpu def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return __a = True __a = model_class(config=lowerCamelCase_ ) __a = self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) __a = torch.jit.trace( lowerCamelCase_ , (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCamelCase_ , os.path.join(lowerCamelCase_ , "bert.pt" ) ) __a = torch.jit.load(os.path.join(lowerCamelCase_ , "bert.pt" ) , map_location=lowerCamelCase_ ) loaded(inputs_dict["input_ids"].to(lowerCamelCase_ ) , inputs_dict["attention_mask"].to(lowerCamelCase_ ) ) @require_torch class snake_case__ ( unittest.TestCase ): @slow def a__ ( self ): __a = NezhaModel.from_pretrained("sijunhe/nezha-cn-base" ) __a = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __a = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ )[0] __a = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , lowerCamelCase_ ) __a = torch.tensor([[[0.0685, 0.2441, 0.1102], [0.0600, 0.1906, 0.1349], [0.0221, 0.0819, 0.0586]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCamelCase_ , atol=1E-4 ) ) @slow def a__ ( self ): __a = NezhaForMaskedLM.from_pretrained("sijunhe/nezha-cn-base" ) __a = torch.tensor([[0, 1, 2, 3, 4, 5]] ) __a = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ )[0] __a = torch.Size((1, 6, 21128) ) self.assertEqual(output.shape , lowerCamelCase_ ) __a = torch.tensor( [[-2.7939, -1.7902, -2.2189], [-2.8585, -1.8908, -2.3723], [-2.6499, -1.7750, -2.2558]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCamelCase_ , atol=1E-4 ) )
718
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Any = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Optional[Any] = { """vocab_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/vocab.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/vocab.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/vocab.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/vocab.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/vocab.json""", }, """merges_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/merges.txt""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/merges.txt""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/merges.txt""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/merges.txt""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/merges.txt""", }, """tokenizer_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/tokenizer.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/tokenizer.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__:Union[str, Any] = { """gpt2""": 1024, """gpt2-medium""": 1024, """gpt2-large""": 1024, """gpt2-xl""": 1024, """distilgpt2""": 1024, } class snake_case__ ( snake_case_ ): _snake_case : Tuple = VOCAB_FILES_NAMES _snake_case : str = PRETRAINED_VOCAB_FILES_MAP _snake_case : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : List[str] = ["""input_ids""", """attention_mask"""] _snake_case : Dict = GPTaTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , add_prefix_space=lowerCamelCase , **lowerCamelCase , ) __a = kwargs.pop("add_bos_token" , lowerCamelCase ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=64 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=[1, 16, 4, 4] , lowerCamelCase=None , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size __a = (self.image_size // 32) ** 2 __a = num_patches + 1 def a__ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, pixel_values, labels def a__ ( self ): __a = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [4, 8, 16, 32], "num_groups": 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=__UpperCamelCase , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTHybridModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() __a = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.type_sequence_label_size __a = ViTHybridForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() __a = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ): __a = self.prepare_config_and_inputs() __a , __a , __a = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( _UpperCAmelCase, _UpperCAmelCase, unittest.TestCase ): _snake_case : Union[str, Any] = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () _snake_case : Dict = ( {'feature-extraction': ViTHybridModel, 'image-classification': ViTHybridForImageClassification} if is_torch_available() else {} ) _snake_case : str = False _snake_case : Dict = False _snake_case : Union[str, Any] = False def a__ ( self ): __a = ViTHybridModelTester(self ) __a = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def a__ ( self ): pass def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(__UpperCamelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() __a = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: __a = model_class(config=__UpperCamelCase ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": __a = [F"{name}.{key}" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) @slow def a__ ( self ): for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ViTHybridModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def _lowerCamelCase( ): __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def a__ ( self ): return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def a__ ( self ): __a = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( __UpperCamelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=__UpperCamelCase , return_tensors="pt" ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): __a = model(**__UpperCamelCase ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) __a = torch.tensor([-1.9090, -0.4993, -0.2389] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow @require_accelerate def a__ ( self ): __a = ViTHybridImageProcessor.from_pretrained("google/vit-hybrid-base-bit-384" ) __a = ViTHybridForImageClassification.from_pretrained("google/vit-hybrid-base-bit-384" , device_map="auto" ) __a = prepare_img() __a = image_processor(images=__UpperCamelCase , return_tensors="pt" ) __a = model(**__UpperCamelCase ) __a = outputs.logits # model predicts one of the 1000 ImageNet classes __a = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , "tabby, tabby cat" )
719
"""simple docstring""" from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("repo_id" , ["canonical_dataset_name", "org-name/dataset-name"] ) @pytest.mark.parametrize("path" , ["filename.csv", "filename with blanks.csv"] ) @pytest.mark.parametrize("revision" , [None, "v2"] ) def _lowerCamelCase( a , a , a ): __a = hf_hub_url(repo_id=a , path=a , revision=a ) assert url == F"https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(a )}"
67
0
"""simple docstring""" import argparse from collections import defaultdict def _lowerCamelCase( a , a , a , a , a ): __a = F"{file}_{class_name}_{test_name}" done_test[_id] += 1 with open(a , "r" ) as f: __a = f.readlines() __a = F"class {class_name}(" __a = F"{4 * ' '}def {test_name}(" __a = F"{8 * ' '}{correct_line.split()[0]}" __a = F"{1_6 * ' '}{correct_line.split()[0]}" __a = False __a = False __a = False __a = False __a = 0 __a = 0 __a = [] for line in lines: if line.startswith(a ): __a = True elif in_class and line.startswith(a ): __a = True elif in_class and in_func and (line.startswith(a ) or line.startswith(a )): __a = len(line.split(correct_line.split()[0] )[0] ) count += 1 if count == done_test[_id]: __a = True if in_class and in_func and in_line: if ")" not in line: continue else: __a = True if in_class and in_func and in_line and insert_line: new_lines.append(F"{spaces * ' '}{correct_line}" ) __a = __a = __a = __a = False else: new_lines.append(a ) with open(a , "w" ) as f: for line in new_lines: f.write(a ) def _lowerCamelCase( a , a=None ): if fail is not None: with open(a , "r" ) as f: __a = {l.strip() for l in f.readlines()} else: __a = None with open(a , "r" ) as f: __a = f.readlines() __a = defaultdict(a ) for line in correct_lines: __a , __a , __a , __a = line.split(";" ) if test_failures is None or "::".join([file, class_name, test_name] ) in test_failures: overwrite_file(a , a , a , a , a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:str = argparse.ArgumentParser() parser.add_argument("""--correct_filename""", help="""filename of tests with expected result""") parser.add_argument("""--fail_filename""", help="""filename of test failures""", type=str, default=None) SCREAMING_SNAKE_CASE__:int = parser.parse_args() main(args.correct_filename, args.fail_filename)
720
"""simple docstring""" from __future__ import annotations def _lowerCamelCase( a , a , a ): if len(a ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(a ) or left < -len(a ) or right >= len(a ) or right < -len(a ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] __a = (left + right) >> 1 # the middle __a = find_max(a , a , a ) # find max in range[left, mid] __a = find_max(a , mid + 1 , a ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
67
0
"""simple docstring""" import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate SCREAMING_SNAKE_CASE__:List[str] = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataRow("""""", """|""", """|"""), datarow=DataRow("""""", """|""", """|"""), padding=1, with_header_hide=None, ) SCREAMING_SNAKE_CASE__:Optional[int] = [] SCREAMING_SNAKE_CASE__:int = [] SCREAMING_SNAKE_CASE__:Optional[int] = {'''type''': '''section''', '''text''': {'''type''': '''plain_text''', '''text''': '''No failed tests! 🤗''', '''emoji''': True}} SCREAMING_SNAKE_CASE__:str = [ { '''type''': '''header''', '''text''': { '''type''': '''plain_text''', '''text''': F'''🤗 Accelerate nightly {os.environ.get('TEST_TYPE', '')} test results''', '''emoji''': True, }, } ] SCREAMING_SNAKE_CASE__:Union[str, Any] = 0 for log in Path().glob("""*.log"""): SCREAMING_SNAKE_CASE__:Dict = 0 with open(log, """r""") as f: for line in f: SCREAMING_SNAKE_CASE__:Dict = json.loads(line) if line.get("""nodeid""", """""") != "": SCREAMING_SNAKE_CASE__:Optional[int] = line['''nodeid'''] if line.get("""duration""", None) is not None: SCREAMING_SNAKE_CASE__:int = F'''{line['duration']:.4f}''' if line.get("""outcome""", """""") == "failed": section_num_failed += 1 failed.append([test, duration, log.name.split("""_""")[0]]) total_num_failed += 1 group_info.append([str(log), section_num_failed, failed]) SCREAMING_SNAKE_CASE__:Dict = [] log.unlink() SCREAMING_SNAKE_CASE__:Union[str, Any] = '''''' SCREAMING_SNAKE_CASE__:Union[str, Any] = [] if total_num_failed > 0: for name, num_failed, failed_tests in group_info: if num_failed > 0: if num_failed == 1: message += F"*{name[1:]}: {num_failed} failed test*\n" else: message += F"*{name[1:]}: {num_failed} failed tests*\n" SCREAMING_SNAKE_CASE__:List[str] = [] SCREAMING_SNAKE_CASE__:Optional[int] = {} for test in failed_tests: SCREAMING_SNAKE_CASE__:Tuple = test[0].split("""::""") SCREAMING_SNAKE_CASE__:Any = data[0].split("""/""")[-1] if data[0] not in filesafailed: SCREAMING_SNAKE_CASE__:Union[str, Any] = [data[1:]] else: filesafailed[data[0]] += [data[1:]] failed_table.append(data) SCREAMING_SNAKE_CASE__:str = [test[0] for test in failed_table] SCREAMING_SNAKE_CASE__:Optional[int] = list(set(files)) # Count number of instances in failed_tests SCREAMING_SNAKE_CASE__:Dict = [] for file in individual_files: table.append([file, len(filesafailed[file])]) SCREAMING_SNAKE_CASE__:List[Any] = tabulate( table, headers=["""Test Location""", """Num Failed"""], tablefmt=hf_table_format, stralign="""right""", ) message += F"\n```\n{failed_table}\n```" all_filesafailed.append(filesafailed) if len(message) > 3000: SCREAMING_SNAKE_CASE__:List[str] = '''Too many failed tests, please see the full report in the Action results.''' SCREAMING_SNAKE_CASE__:List[Any] = len(err) + 10 SCREAMING_SNAKE_CASE__:int = message[: 3000 - offset] + F'''\n...\n```\n{err}''' print(F'''### {message}''') else: SCREAMING_SNAKE_CASE__:Optional[Any] = '''No failed tests! 🤗''' print(F'''## {message}''') payload.append(no_error_payload) if os.environ.get("""TEST_TYPE""", """""") != "": from slack_sdk import WebClient SCREAMING_SNAKE_CASE__:List[str] = WebClient(token=os.environ["""SLACK_API_TOKEN"""]) if message != "No failed tests! 🤗": SCREAMING_SNAKE_CASE__:List[str] = { '''type''': '''section''', '''text''': { '''type''': '''mrkdwn''', '''text''': message, }, } payload.append(md_report) SCREAMING_SNAKE_CASE__:Optional[int] = { '''type''': '''section''', '''text''': { '''type''': '''mrkdwn''', '''text''': '''*For more details:*''', }, '''accessory''': { '''type''': '''button''', '''text''': { '''type''': '''plain_text''', '''text''': '''Check Action results''', '''emoji''': True, }, '''url''': F'''https://github.com/{os.environ['GITHUB_REPOSITORY']}/actions/runs/{os.environ['GITHUB_RUN_ID']}''', }, } payload.append(action_button) SCREAMING_SNAKE_CASE__:List[str] = { '''type''': '''context''', '''elements''': [ { '''type''': '''plain_text''', '''text''': F'''Nightly {os.environ.get('TEST_TYPE')} test results for {date.today()}''', } ], } payload.append(date_report) SCREAMING_SNAKE_CASE__:Optional[Any] = client.chat_postMessage(channel="""#accelerate-ci-daily""", text=message, blocks=payload) SCREAMING_SNAKE_CASE__:Union[str, Any] = response.data['''ts'''] for failed_file in all_filesafailed: for test_location, test_failures in failed_file.items(): # Keep only the first instance of the test name SCREAMING_SNAKE_CASE__:int = '''''' for i, row in enumerate(test_failures): if row[0] != test_class: SCREAMING_SNAKE_CASE__:List[str] = row[0] else: SCREAMING_SNAKE_CASE__:Optional[Any] = '''''' SCREAMING_SNAKE_CASE__:Union[str, Any] = { '''type''': '''section''', '''text''': { '''type''': '''mrkdwn''', '''text''': F'''Test location: {test_location}\n```\n{tabulate(test_failures, headers=['Class', 'Test'], tablefmt=hf_table_format, stralign='right')}\n```''', }, } client.chat_postMessage( channel="""#accelerate-ci-daily""", thread_ts=ts, blocks=[payload], )
721
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json""", """google/bigbird-roberta-large""": """https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json""", """google/bigbird-base-trivia-itc""": """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json""", # See all BigBird models at https://huggingface.co/models?filter=big_bird } class snake_case__ ( snake_case_ ): _snake_case : Any = """big_bird""" def __init__( self , lowerCamelCase=50358 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu_new" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=4096 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=True , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=66 , lowerCamelCase="block_sparse" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=64 , lowerCamelCase=3 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , sep_token_id=lowerCamelCase , **lowerCamelCase , ) __a = vocab_size __a = max_position_embeddings __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = initializer_range __a = type_vocab_size __a = layer_norm_eps __a = use_cache __a = rescale_embeddings __a = attention_type __a = use_bias __a = block_size __a = num_random_blocks __a = classifier_dropout class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task == "multiple-choice": __a = {0: "batch", 1: "choice", 2: "sequence"} else: __a = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
67
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Any = {"""vocab_file""": """spm_char.model"""} SCREAMING_SNAKE_CASE__:Dict = { """vocab_file""": { """microsoft/speecht5_asr""": """https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model""", """microsoft/speecht5_tts""": """https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model""", """microsoft/speecht5_vc""": """https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model""", } } SCREAMING_SNAKE_CASE__:List[str] = { """microsoft/speecht5_asr""": 1024, """microsoft/speecht5_tts""": 1024, """microsoft/speecht5_vc""": 1024, } class snake_case__ ( _lowercase ): _snake_case : Optional[Any] = VOCAB_FILES_NAMES _snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : Tuple = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCamelCase , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="<unk>" , lowerCamelCase="<pad>" , lowerCamelCase = None , **lowerCamelCase , ): __a = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=A_ , eos_token=A_ , unk_token=A_ , pad_token=A_ , sp_model_kwargs=self.sp_model_kwargs , **A_ , ) __a = vocab_file __a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) @property def a__ ( self ): return self.sp_model.get_piece_size() def a__ ( self ): __a = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __a = self.__dict__.copy() __a = None return state def __setstate__( self , lowerCamelCase ): __a = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __a = {} __a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def a__ ( self , lowerCamelCase ): return self.sp_model.encode(A_ , out_type=A_ ) def a__ ( self , lowerCamelCase ): return self.sp_model.piece_to_id(A_ ) def a__ ( self , lowerCamelCase ): __a = self.sp_model.IdToPiece(A_ ) return token def a__ ( self , lowerCamelCase ): __a = [] __a = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(A_ ) + token __a = [] else: current_sub_tokens.append(A_ ) out_string += self.sp_model.decode(A_ ) return out_string.strip() def a__ ( self , lowerCamelCase , lowerCamelCase=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A_ , token_ids_a=A_ , already_has_special_tokens=A_ ) __a = [1] if token_ids_a is None: return ([0] * len(A_ )) + suffix_ones return ([0] * len(A_ )) + ([0] * len(A_ )) + suffix_ones def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if not os.path.isdir(A_ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __a = os.path.join( A_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ , "wb" ) as fi: __a = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,)
700
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:Optional[int] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[int] = {"""tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Tuple = { """tokenizer_file""": { """bigscience/tokenizer""": """https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json""", """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json""", }, } class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = VOCAB_FILES_NAMES _snake_case : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Optional[int] = ["""input_ids""", """attention_mask"""] _snake_case : Optional[int] = None def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<unk>" , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="<pad>" , lowerCamelCase=False , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , pad_token=lowerCamelCase , add_prefix_space=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , **lowerCamelCase , ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
"""simple docstring""" import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa SCREAMING_SNAKE_CASE__:Any = logging.getLogger(__name__) class snake_case__ ( lowercase__ ): _snake_case : Union[str, Any] = 'summarization' _snake_case : Any = ['loss'] _snake_case : List[str] = ROUGE_KEYS _snake_case : Dict = 'rouge2' def __init__( self , lowerCamelCase , **lowerCamelCase ): if hparams.sortish_sampler and hparams.gpus > 1: __a = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError("Dynamic Batch size does not work for multi-gpu training" ) if hparams.sortish_sampler: raise ValueError("--sortish_sampler and --max_tokens_per_batch may not be used simultaneously" ) super().__init__(lowerCamelCase , num_labels=lowerCamelCase , mode=self.mode , **lowerCamelCase ) use_task_specific_params(self.model , "summarization" ) save_git_info(self.hparams.output_dir ) __a = Path(self.output_dir ) / "metrics.json" __a = Path(self.output_dir ) / "hparams.pkl" pickle_save(self.hparams , self.hparams_save_path ) __a = 0 __a = defaultdict(lowerCamelCase ) __a = self.config.model_type __a = self.config.tgt_vocab_size if self.model_type == "fsmt" else self.config.vocab_size __a = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } __a = { "train": self.hparams.n_train, "val": self.hparams.n_val, "test": self.hparams.n_test, } __a = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} __a = { "train": self.hparams.max_target_length, "val": self.hparams.val_max_target_length, "test": self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], F"target_lens: {self.target_lens}" assert self.target_lens["train"] <= self.target_lens["test"], F"target_lens: {self.target_lens}" if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) __a = get_git_info()["repo_sha"] __a = hparams.num_workers __a = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , lowerCamelCase ): __a = self.tokenizer.lang_code_to_id[hparams.tgt_lang] __a = self.decoder_start_token_id __a = ( SeqaSeqDataset if hasattr(self.tokenizer , "prepare_seq2seq_batch" ) else LegacySeqaSeqDataset ) __a = False __a = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: __a = self.hparams.eval_max_gen_length else: __a = self.model.config.max_length __a = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def a__ ( self , lowerCamelCase ): __a = { k: self.tokenizer.batch_decode(v.tolist() ) if "mask" not in k else v.shape for k, v in batch.items() } save_json(lowerCamelCase , Path(self.output_dir ) / "text_batch.json" ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / "tok_batch.json" ) __a = True return readable_batch def a__ ( self , lowerCamelCase , **lowerCamelCase ): return self.model(lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = self.tokenizer.batch_decode( lowerCamelCase , skip_special_tokens=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase ) return lmap(str.strip , lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = self.tokenizer.pad_token_id __a , __a = batch["input_ids"], batch["attention_mask"] __a = batch["labels"] if isinstance(self.model , lowerCamelCase ): __a = self.model._shift_right(lowerCamelCase ) else: __a = shift_tokens_right(lowerCamelCase , lowerCamelCase ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero __a = decoder_input_ids self.save_readable_batch(lowerCamelCase ) __a = self(lowerCamelCase , attention_mask=lowerCamelCase , decoder_input_ids=lowerCamelCase , use_cache=lowerCamelCase ) __a = outputs["logits"] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id __a = nn.CrossEntropyLoss(ignore_index=lowerCamelCase ) assert lm_logits.shape[-1] == self.vocab_size __a = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: __a = nn.functional.log_softmax(lowerCamelCase , dim=-1 ) __a , __a = label_smoothed_nll_loss( lowerCamelCase , lowerCamelCase , self.hparams.label_smoothing , ignore_index=lowerCamelCase ) return (loss,) @property def a__ ( self ): return self.tokenizer.pad_token_id def a__ ( self , lowerCamelCase , lowerCamelCase ): __a = self._step(lowerCamelCase ) __a = dict(zip(self.loss_names , lowerCamelCase ) ) # tokens per batch __a = batch["input_ids"].ne(self.pad ).sum() + batch["labels"].ne(self.pad ).sum() __a = batch["input_ids"].shape[0] __a = batch["input_ids"].eq(self.pad ).sum() __a = batch["input_ids"].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def a__ ( self , lowerCamelCase , lowerCamelCase ): return self._generative_step(lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase="val" ): self.step_count += 1 __a = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} __a = losses["loss"] __a = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ["gen_time", "gen_len"] } __a = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) __a = torch.tensor(lowerCamelCase ).type_as(lowerCamelCase ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(lowerCamelCase ) __a = {F"{prefix}_avg_{k}": x for k, x in losses.items()} __a = self.step_count self.metrics[prefix].append(lowerCamelCase ) # callback writes this to self.metrics_save_path __a = flatten_list([x["preds"] for x in outputs] ) return { "log": all_metrics, "preds": preds, F"{prefix}_loss": loss, F"{prefix}_{self.val_metric}": metric_tensor, } def a__ ( self , lowerCamelCase , lowerCamelCase ): return calculate_rouge(lowerCamelCase , lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') __a = self.model.generate( batch["input_ids"] , attention_mask=batch["attention_mask"] , use_cache=lowerCamelCase , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) __a = (time.time() - ta) / batch["input_ids"].shape[0] __a = self.ids_to_clean_text(lowerCamelCase ) __a = self.ids_to_clean_text(batch["labels"] ) __a = self._step(lowerCamelCase ) __a = dict(zip(self.loss_names , lowerCamelCase ) ) __a = self.calc_generative_metrics(lowerCamelCase , lowerCamelCase ) __a = np.mean(lmap(lowerCamelCase , lowerCamelCase ) ) base_metrics.update(gen_time=lowerCamelCase , gen_len=lowerCamelCase , preds=lowerCamelCase , target=lowerCamelCase , **lowerCamelCase ) return base_metrics def a__ ( self , lowerCamelCase , lowerCamelCase ): return self._generative_step(lowerCamelCase ) def a__ ( self , lowerCamelCase ): return self.validation_epoch_end(lowerCamelCase , prefix="test" ) def a__ ( self , lowerCamelCase ): __a = self.n_obs[type_path] __a = self.target_lens[type_path] __a = self.dataset_class( self.tokenizer , type_path=lowerCamelCase , n_obs=lowerCamelCase , max_target_length=lowerCamelCase , **self.dataset_kwargs , ) return dataset def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = False ): __a = self.get_dataset(lowerCamelCase ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": __a = dataset.make_sortish_sampler(lowerCamelCase , distributed=self.hparams.gpus > 1 ) return DataLoader( lowerCamelCase , batch_size=lowerCamelCase , collate_fn=dataset.collate_fn , shuffle=lowerCamelCase , num_workers=self.num_workers , sampler=lowerCamelCase , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": __a = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( lowerCamelCase , batch_sampler=lowerCamelCase , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( lowerCamelCase , batch_size=lowerCamelCase , collate_fn=dataset.collate_fn , shuffle=lowerCamelCase , num_workers=self.num_workers , sampler=lowerCamelCase , ) def a__ ( self ): __a = self.get_dataloader("train" , batch_size=self.hparams.train_batch_size , shuffle=lowerCamelCase ) return dataloader def a__ ( self ): return self.get_dataloader("val" , batch_size=self.hparams.eval_batch_size ) def a__ ( self ): return self.get_dataloader("test" , batch_size=self.hparams.eval_batch_size ) @staticmethod def a__ ( lowerCamelCase , lowerCamelCase ): BaseTransformer.add_model_specific_args(lowerCamelCase , lowerCamelCase ) add_generic_args(lowerCamelCase , lowerCamelCase ) parser.add_argument( "--max_source_length" , default=1024 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--max_target_length" , default=56 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--val_max_target_length" , default=142 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--test_max_target_length" , default=142 , type=lowerCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument("--freeze_encoder" , action="store_true" ) parser.add_argument("--freeze_embeds" , action="store_true" ) parser.add_argument("--sortish_sampler" , action="store_true" , default=lowerCamelCase ) parser.add_argument("--overwrite_output_dir" , action="store_true" , default=lowerCamelCase ) parser.add_argument("--max_tokens_per_batch" , type=lowerCamelCase , default=lowerCamelCase ) parser.add_argument("--logger_name" , type=lowerCamelCase , choices=["default", "wandb", "wandb_shared"] , default="default" ) parser.add_argument("--n_train" , type=lowerCamelCase , default=-1 , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--n_val" , type=lowerCamelCase , default=500 , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--n_test" , type=lowerCamelCase , default=-1 , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument( "--task" , type=lowerCamelCase , default="summarization" , required=lowerCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--label_smoothing" , type=lowerCamelCase , default=0.0 , required=lowerCamelCase ) parser.add_argument("--src_lang" , type=lowerCamelCase , default="" , required=lowerCamelCase ) parser.add_argument("--tgt_lang" , type=lowerCamelCase , default="" , required=lowerCamelCase ) parser.add_argument("--eval_beams" , type=lowerCamelCase , default=lowerCamelCase , required=lowerCamelCase ) parser.add_argument( "--val_metric" , type=lowerCamelCase , default=lowerCamelCase , required=lowerCamelCase , choices=["bleu", "rouge2", "loss", None] ) parser.add_argument("--eval_max_gen_length" , type=lowerCamelCase , default=lowerCamelCase , help="never generate more than n tokens" ) parser.add_argument("--save_top_k" , type=lowerCamelCase , default=1 , required=lowerCamelCase , help="How many checkpoints to save" ) parser.add_argument( "--early_stopping_patience" , type=lowerCamelCase , default=-1 , required=lowerCamelCase , help=( "-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So" " val_check_interval will effect it." ) , ) return parser class snake_case__ ( lowercase__ ): _snake_case : Any = 'translation' _snake_case : Optional[Any] = ['loss'] _snake_case : Optional[Any] = ['bleu'] _snake_case : List[str] = 'bleu' def __init__( self , lowerCamelCase , **lowerCamelCase ): super().__init__(lowerCamelCase , **lowerCamelCase ) __a = hparams.src_lang __a = hparams.tgt_lang def a__ ( self , lowerCamelCase , lowerCamelCase ): return calculate_bleu(lowerCamelCase , lowerCamelCase ) def _lowerCamelCase( a , a=None ): Path(args.output_dir ).mkdir(exist_ok=__SCREAMING_SNAKE_CASE ) check_output_dir(__SCREAMING_SNAKE_CASE , expected_items=3 ) if model is None: if "summarization" in args.task: __a = SummarizationModule(__SCREAMING_SNAKE_CASE ) else: __a = TranslationModule(__SCREAMING_SNAKE_CASE ) __a = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith("/tmp" ) or str(args.output_dir ).startswith("/var" ) ): __a = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger __a = os.environ.get("WANDB_PROJECT" , __SCREAMING_SNAKE_CASE ) __a = WandbLogger(name=model.output_dir.name , project=__SCREAMING_SNAKE_CASE ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger __a = WandbLogger(name=model.output_dir.name , project=F"hf_{dataset}" ) if args.early_stopping_patience >= 0: __a = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: __a = False __a = args.val_metric == "loss" __a = generic_train( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , __SCREAMING_SNAKE_CASE ) , early_stopping_callback=__SCREAMING_SNAKE_CASE , logger=__SCREAMING_SNAKE_CASE , ) pickle_save(model.hparams , model.output_dir / "hparams.pkl" ) if not args.do_predict: return model __a = "" __a = sorted(glob.glob(os.path.join(args.output_dir , "*.ckpt" ) , recursive=__SCREAMING_SNAKE_CASE ) ) if checkpoints: __a = checkpoints[-1] __a = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": SCREAMING_SNAKE_CASE__:str = argparse.ArgumentParser() SCREAMING_SNAKE_CASE__:str = pl.Trainer.add_argparse_args(parser) SCREAMING_SNAKE_CASE__:Tuple = SummarizationModule.add_model_specific_args(parser, os.getcwd()) SCREAMING_SNAKE_CASE__:Optional[int] = parser.parse_args() main(args)
701
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class snake_case__ : _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : int _snake_case : int _snake_case : float _snake_case : float _snake_case : Tuple[int] def a__ ( self ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def a__ ( self ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def a__ ( self ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def a__ ( self ): __a = torch.arange(self.height * self.width ) __a = torch.stack( [ pixel_indices % self.width, torch.div(lowerCamelCase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def a__ ( self ): __a , *__a = self.shape __a = int(np.prod(lowerCamelCase ) ) __a = self.get_image_coords() __a = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __a = self.get_camera_rays(lowerCamelCase ) __a = rays.view(lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def a__ ( self , lowerCamelCase ): __a , *__a , __a = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __a = coords.view(lowerCamelCase , -1 , 2 ) __a = self.resolution() __a = self.fov() __a = (flat.float() / (res - 1)) * 2 - 1 __a = fracs * torch.tan(fov / 2 ) __a = fracs.view(lowerCamelCase , -1 , 2 ) __a = ( self.z.view(lowerCamelCase , 1 , 3 ) + self.x.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) __a = directions / directions.norm(dim=-1 , keepdim=lowerCamelCase ) __a = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCamelCase , *lowerCamelCase , 2 , 3 ) def a__ ( self , lowerCamelCase , lowerCamelCase ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCamelCase , height=lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowerCamelCase( a ): __a = [] __a = [] __a = [] __a = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __a = np.array([np.sin(a ), np.cos(a ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __a = -z * 4 __a = np.array([np.cos(a ), -np.sin(a ), 0.0] ) __a = np.cross(a , a ) origins.append(a ) xs.append(a ) ys.append(a ) zs.append(a ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(a , axis=0 ) ).float() , x=torch.from_numpy(np.stack(a , axis=0 ) ).float() , y=torch.from_numpy(np.stack(a , axis=0 ) ).float() , z=torch.from_numpy(np.stack(a , axis=0 ) ).float() , width=a , height=a , x_fov=0.7 , y_fov=0.7 , shape=(1, len(a )) , )
67
0
"""simple docstring""" from math import factorial def _lowerCamelCase( a = 1_0_0 ): return sum(map(a , str(factorial(a ) ) ) ) if __name__ == "__main__": print(solution(int(input("""Enter the Number: """).strip())))
702
"""simple docstring""" def _lowerCamelCase( a ): return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase( a ): __a = 0 __a = number while duplicate > 0: __a , __a = divmod(a , 1_0 ) fact_sum += factorial(a ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__:Optional[Any] = int(input("""Enter number: """).strip()) print( F'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
67
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__:Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Dict = { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/config.json""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/config.json""", """funnel-transformer/medium-base""": """https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json""", """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/config.json""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json""", """funnel-transformer/xlarge-base""": """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json""", } class snake_case__ ( UpperCamelCase_ ): _snake_case : Any = """funnel""" _snake_case : List[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", } def __init__( self , lowerCamelCase=30522 , lowerCamelCase=[4, 4, 4] , lowerCamelCase=None , lowerCamelCase=2 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=64 , lowerCamelCase=3072 , lowerCamelCase="gelu_new" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=0.1 , lowerCamelCase=None , lowerCamelCase=1E-9 , lowerCamelCase="mean" , lowerCamelCase="relative_shift" , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , **lowerCamelCase , ): __a = vocab_size __a = block_sizes __a = [1] * len(__A ) if block_repeats is None else block_repeats assert len(__A ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." __a = num_decoder_layers __a = d_model __a = n_head __a = d_head __a = d_inner __a = hidden_act __a = hidden_dropout __a = attention_dropout __a = activation_dropout __a = initializer_range __a = initializer_std __a = layer_norm_eps assert pooling_type in [ "mean", "max", ], F"Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported." __a = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F"Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported." __a = attention_type __a = separate_cls __a = truncate_seq __a = pool_q_only super().__init__(**__A ) @property def a__ ( self ): return sum(self.block_sizes ) @num_hidden_layers.setter def a__ ( self , lowerCamelCase ): raise NotImplementedError( "This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`." ) @property def a__ ( self ): return len(self.block_sizes ) @num_blocks.setter def a__ ( self , lowerCamelCase ): raise NotImplementedError("This model does not support the setting of `num_blocks`. Please set `block_sizes`." )
703
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE__:Optional[Any] = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Union[str, Any] = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def _lowerCamelCase( a , a ): __a = old_name if "patch_embed" in old_name: __a = old_name.split("." ) if layer == "0": __a = old_name.replace("0" , "convolution1" ) elif layer == "1": __a = old_name.replace("1" , "batchnorm_before" ) elif layer == "3": __a = old_name.replace("3" , "convolution2" ) else: __a = old_name.replace("4" , "batchnorm_after" ) if "network" in old_name and re.search(R"\d\.\d" , _lowerCamelCase ): __a = r"\b\d{2}\b" if bool(re.search(_lowerCamelCase , _lowerCamelCase ) ): __a = re.search(R"\d\.\d\d." , _lowerCamelCase ).group() else: __a = re.search(R"\d\.\d." , _lowerCamelCase ).group() if int(match[0] ) < 6: __a = old_name.replace(_lowerCamelCase , "" ) __a = trimmed_name.replace("network" , match[0] + ".meta4D_layers.blocks." + match[2:-1] ) __a = "intermediate_stages." + trimmed_name else: __a = old_name.replace(_lowerCamelCase , "" ) if int(match[2] ) < num_meta4D_last_stage: __a = trimmed_name.replace("network" , "meta4D_layers.blocks." + match[2] ) else: __a = str(int(match[2] ) - num_meta4D_last_stage ) __a = trimmed_name.replace("network" , "meta3D_layers.blocks." + layer_index ) if "norm1" in old_name: __a = trimmed_name.replace("norm1" , "layernorm1" ) elif "norm2" in old_name: __a = trimmed_name.replace("norm2" , "layernorm2" ) elif "fc1" in old_name: __a = trimmed_name.replace("fc1" , "linear_in" ) elif "fc2" in old_name: __a = trimmed_name.replace("fc2" , "linear_out" ) __a = "last_stage." + trimmed_name elif "network" in old_name and re.search(R".\d." , _lowerCamelCase ): __a = old_name.replace("network" , "intermediate_stages" ) if "fc" in new_name: __a = new_name.replace("fc" , "convolution" ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): __a = new_name.replace("norm1" , "batchnorm_before" ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): __a = new_name.replace("norm2" , "batchnorm_after" ) if "proj" in new_name: __a = new_name.replace("proj" , "projection" ) if "dist_head" in new_name: __a = new_name.replace("dist_head" , "distillation_classifier" ) elif "head" in new_name: __a = new_name.replace("head" , "classifier" ) elif "patch_embed" in new_name: __a = "efficientformer." + new_name elif new_name == "norm.weight" or new_name == "norm.bias": __a = new_name.replace("norm" , "layernorm" ) __a = "efficientformer." + new_name else: __a = "efficientformer.encoder." + new_name return new_name def _lowerCamelCase( a , a ): for key in checkpoint.copy().keys(): __a = checkpoint.pop(_lowerCamelCase ) __a = val return checkpoint def _lowerCamelCase( ): __a = "http://images.cocodataset.org/val2017/000000039769.jpg" __a = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase ).raw ) return image def _lowerCamelCase( a , a , a , a ): __a = torch.load(_lowerCamelCase , map_location="cpu" )["model"] __a = EfficientFormerConfig.from_json_file(_lowerCamelCase ) __a = EfficientFormerForImageClassificationWithTeacher(_lowerCamelCase ) __a = "_".join(checkpoint_path.split("/" )[-1].split("." )[0].split("_" )[:-1] ) __a = config.depths[-1] - config.num_metaad_blocks + 1 __a = convert_torch_checkpoint(_lowerCamelCase , _lowerCamelCase ) model.load_state_dict(_lowerCamelCase ) model.eval() __a = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } # prepare image __a = prepare_img() __a = 2_5_6 __a = 2_2_4 __a = EfficientFormerImageProcessor( size={"shortest_edge": image_size} , crop_size={"height": crop_size, "width": crop_size} , resample=pillow_resamplings["bicubic"] , ) __a = processor(images=_lowerCamelCase , return_tensors="pt" ).pixel_values # original processing pipeline __a = Compose( [ Resize(_lowerCamelCase , interpolation=pillow_resamplings["bicubic"] ), CenterCrop(_lowerCamelCase ), ToTensor(), Normalize(_lowerCamelCase , _lowerCamelCase ), ] ) __a = image_transforms(_lowerCamelCase ).unsqueeze(0 ) assert torch.allclose(_lowerCamelCase , _lowerCamelCase ) __a = model(_lowerCamelCase ) __a = outputs.logits __a = (1, 1_0_0_0) if "l1" in model_name: __a = torch.Tensor( [-0.13_12, 0.43_53, -1.04_99, -0.51_24, 0.41_83, -0.67_93, -1.37_77, -0.08_93, -0.73_58, -2.43_28] ) assert torch.allclose(logits[0, :1_0] , _lowerCamelCase , atol=1E-3 ) assert logits.shape == expected_shape elif "l3" in model_name: __a = torch.Tensor( [-1.31_50, -1.54_56, -1.25_56, -0.84_96, -0.71_27, -0.78_97, -0.97_28, -0.30_52, 0.37_51, -0.31_27] ) assert torch.allclose(logits[0, :1_0] , _lowerCamelCase , atol=1E-3 ) assert logits.shape == expected_shape elif "l7" in model_name: __a = torch.Tensor( [-1.02_83, -1.41_31, -0.56_44, -1.31_15, -0.57_85, -1.20_49, -0.75_28, 0.19_92, -0.38_22, -0.08_78] ) assert logits.shape == expected_shape else: raise ValueError( F"Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7" ) # Save Checkpoints Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) model.save_pretrained(_lowerCamelCase ) print(F"Checkpoint successfuly converted. Model saved at {pytorch_dump_path}" ) processor.save_pretrained(_lowerCamelCase ) print(F"Processor successfuly saved at {pytorch_dump_path}" ) if push_to_hub: print("Pushing model to the hub..." ) model.push_to_hub( repo_id=F"Bearnardd/{pytorch_dump_path}" , commit_message="Add model" , use_temp_dir=_lowerCamelCase , ) processor.push_to_hub( repo_id=F"Bearnardd/{pytorch_dump_path}" , commit_message="Add image processor" , use_temp_dir=_lowerCamelCase , ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:int = argparse.ArgumentParser() # Required parameters parser.add_argument( """--pytorch_model_path""", default=None, type=str, required=True, help="""Path to EfficientFormer pytorch checkpoint.""", ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The json file for EfficientFormer model config.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") parser.add_argument( """--no-push_to_hub""", dest="""push_to_hub""", action="""store_false""", help="""Do not push model and image processor to the hub""", ) parser.set_defaults(push_to_hub=True) SCREAMING_SNAKE_CASE__:List[Any] = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
704
"""simple docstring""" import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def _lowerCamelCase( a , a , a ): __a = OmegaConf.load(a ) __a = torch.load(a , map_location="cpu" )["model"] __a = list(state_dict.keys() ) # extract state_dict for VQVAE __a = {} __a = "first_stage_model." for key in keys: if key.startswith(a ): __a = state_dict[key] # extract state_dict for UNetLDM __a = {} __a = "model.diffusion_model." for key in keys: if key.startswith(a ): __a = state_dict[key] __a = config.model.params.first_stage_config.params __a = config.model.params.unet_config.params __a = VQModel(**a ).eval() vqvae.load_state_dict(a ) __a = UNetLDMModel(**a ).eval() unet.load_state_dict(a ) __a = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=a , ) __a = LDMPipeline(a , a , a ) pipeline.save_pretrained(a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
67
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=3 , lowerCamelCase=224 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=[0.5, 0.5, 0.5] , ): __a = size if size is not None else {"""height""": 18, """width""": 18} __a = parent __a = batch_size __a = num_channels __a = image_size __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std def a__ ( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class snake_case__ ( __SCREAMING_SNAKE_CASE, unittest.TestCase ): _snake_case : Any = ViTImageProcessor if is_vision_available() else None def a__ ( self ): __a = EfficientFormerImageProcessorTester(self ) @property def a__ ( self ): return self.image_proc_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , "image_mean" ) ) self.assertTrue(hasattr(_a , "image_std" ) ) self.assertTrue(hasattr(_a , "do_normalize" ) ) self.assertTrue(hasattr(_a , "do_resize" ) ) self.assertTrue(hasattr(_a , "size" ) ) def a__ ( self ): pass def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_proc_tester , equal_resolution=_a ) for image in image_inputs: self.assertIsInstance(_a , Image.Image ) # Test not batched input __a = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __a = image_processor(_a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_proc_tester , equal_resolution=_a , numpify=_a ) for image in image_inputs: self.assertIsInstance(_a , np.ndarray ) # Test not batched input __a = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __a = image_processor(_a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_proc_tester , equal_resolution=_a , torchify=_a ) for image in image_inputs: self.assertIsInstance(_a , torch.Tensor ) # Test not batched input __a = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __a = image_processor(_a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
705
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[Any] = { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json""", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case__ ( snake_case_ ): _snake_case : str = """blenderbot-small""" _snake_case : str = ["""past_key_values"""] _snake_case : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCamelCase=50265 , lowerCamelCase=512 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="gelu" , lowerCamelCase=512 , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , lowerCamelCase=1 , lowerCamelCase=False , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=2 , **lowerCamelCase , ): __a = vocab_size __a = max_position_embeddings __a = d_model __a = encoder_ffn_dim __a = encoder_layers __a = encoder_attention_heads __a = decoder_ffn_dim __a = decoder_layers __a = decoder_attention_heads __a = dropout __a = attention_dropout __a = activation_dropout __a = activation_function __a = init_std __a = encoder_layerdrop __a = decoder_layerdrop __a = use_cache __a = encoder_layers __a = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , is_encoder_decoder=lowerCamelCase , decoder_start_token_id=lowerCamelCase , forced_eos_token_id=lowerCamelCase , **lowerCamelCase , ) class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a = {0: "batch"} __a = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __a = {0: "batch", 1: "decoder_sequence"} __a = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} else: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = super().outputs else: __a = super(lowerCamelCase , self ).outputs if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Generate decoder inputs __a = seq_length if not self.use_past else 1 __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) __a = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} __a = dict(**lowerCamelCase , **lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape __a = common_inputs["decoder_input_ids"].shape[1] __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = decoder_seq_length + 3 __a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __a = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase )] , dim=1 ) __a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __a , __a = self.num_layers __a = min(lowerCamelCase , lowerCamelCase ) __a = max(lowerCamelCase , lowerCamelCase ) - min_num_layers __a = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), ) ) # TODO: test this. __a = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(lowerCamelCase , lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape # Not using the same length for past_key_values __a = seqlen + 2 __a , __a = self.num_layers __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = common_inputs["attention_mask"].dtype __a = torch.cat( [common_inputs["attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) __a = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(lowerCamelCase ) ] return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __a = tokenizer.num_special_tokens_to_add(lowerCamelCase ) __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence __a = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __a = dict(tokenizer(lowerCamelCase , return_tensors=lowerCamelCase ) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): if self.task in ["default", "seq2seq-lm"]: __a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) elif self.task == "causal-lm": __a = self._generate_dummy_inputs_for_causal_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) else: __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if self.task in ["default", "seq2seq-lm"]: __a = super()._flatten_past_key_values_(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __a = super(lowerCamelCase , self )._flatten_past_key_values_( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device SCREAMING_SNAKE_CASE__:List[str] = False class snake_case__ ( unittest.TestCase ): pass @nightly @require_torch_gpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) __a = torch.manual_seed(0 ) __a = pipe.dual_guided( prompt="first prompt" , image=a_ , text_to_image_strength=0.75 , generator=a_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(a_ ) __a = VersatileDiffusionPipeline.from_pretrained(a_ , torch_dtype=torch.floataa ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __a = generator.manual_seed(0 ) __a = pipe.dual_guided( prompt="first prompt" , image=a_ , text_to_image_strength=0.75 , generator=a_ , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def a__ ( self ): __a = VersatileDiffusionPipeline.from_pretrained("shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __a = """cyberpunk 2077""" __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) __a = torch.manual_seed(0 ) __a = pipe.dual_guided( prompt=a_ , image=a_ , text_to_image_strength=0.75 , generator=a_ , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images __a = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __a = np.array([0.1448, 0.1619, 0.1741, 0.1086, 0.1147, 0.1128, 0.1199, 0.1165, 0.1001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __a = """A painting of a squirrel eating a burger """ __a = torch.manual_seed(0 ) __a = pipe.text_to_image( prompt=a_ , generator=a_ , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images __a = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __a = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 __a = pipe.image_variation(a_ , generator=a_ , output_type="numpy" ).images __a = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __a = np.array([0.3076, 0.3123, 0.3284, 0.3782, 0.3770, 0.3894, 0.4297, 0.4331, 0.4456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
706
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=99 , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=9 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase=8 , lowerCamelCase=0.1 , lowerCamelCase=0.002 , lowerCamelCase=1 , lowerCamelCase=0 , lowerCamelCase=0 , lowerCamelCase=None , lowerCamelCase=None , ): __a = parent __a = batch_size __a = encoder_seq_length __a = decoder_seq_length # For common tests __a = self.decoder_seq_length __a = is_training __a = use_attention_mask __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = d_ff __a = relative_attention_num_buckets __a = dropout_rate __a = initializer_factor __a = eos_token_id __a = pad_token_id __a = decoder_start_token_id __a = None __a = decoder_layers def a__ ( self ): return TaConfig.from_pretrained("google/umt5-base" ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): if attention_mask is None: __a = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __a = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __a = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=lowerCamelCase ) if decoder_head_mask is None: __a = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) if cross_attn_head_mask is None: __a = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def a__ ( self ): __a = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __a = input_ids.clamp(self.pad_token_id + 1 ) __a = decoder_input_ids.clamp(self.pad_token_id + 1 ) __a = self.get_config() __a = config.num_attention_heads __a = self.prepare_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, input_dict def a__ ( self ): __a , __a = self.prepare_config_and_inputs() return config, inputs_dict def a__ ( self ): return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self ): return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase , attention_mask=lowerCamelCase , decoder_attention_mask=lowerCamelCase , ) __a = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ) __a = result.last_hidden_state __a = result.past_key_values __a = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).get_decoder().to(lowerCamelCase ).eval() # first forward pass __a = model(lowerCamelCase , use_cache=lowerCamelCase ) __a = model(lowerCamelCase ) __a = model(lowerCamelCase , use_cache=lowerCamelCase ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) + 1 ) __a , __a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = model(lowerCamelCase )["last_hidden_state"] __a = model(lowerCamelCase , past_key_values=lowerCamelCase )["last_hidden_state"] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -1, random_slice_idx].detach() __a = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).to(lowerCamelCase ).half().eval() __a = model(**lowerCamelCase )["last_hidden_state"] self.parent.assertFalse(torch.isnan(lowerCamelCase ).any().item() ) @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _snake_case : int = (UMTaForConditionalGeneration,) if is_torch_available() else () _snake_case : Optional[int] = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _snake_case : List[Any] = True _snake_case : Union[str, Any] = False _snake_case : Union[str, Any] = False _snake_case : Tuple = True _snake_case : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests _snake_case : Optional[Any] = [0.8, 0.9] def a__ ( self ): __a = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() __a = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=lowerCamelCase , opset_version=9 , input_names=["input_ids", "decoder_input_ids"] , ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase ) def a__ ( self ): __a = ["encoder_attentions", "decoder_attentions", "cross_attentions"] __a = self.model_tester.prepare_config_and_inputs() __a = config_and_inputs[0] __a = UMTaForConditionalGeneration(lowerCamelCase ).eval() model.to(lowerCamelCase ) __a = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=lowerCamelCase ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), } for attn_name, (name, mask) in zip(lowerCamelCase , head_masking.items() ): __a = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": __a = torch.ones( config.num_decoder_layers , config.num_heads , device=lowerCamelCase ) __a = model.generate( config_and_inputs[1]["input_ids"] , num_beams=1 , max_length=3 , output_attentions=lowerCamelCase , return_dict_in_generate=lowerCamelCase , **lowerCamelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step __a = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def a__ ( self ): pass @require_torch @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def a__ ( self ): __a = UMTaForConditionalGeneration.from_pretrained("google/umt5-small" , return_dict=lowerCamelCase ).to(lowerCamelCase ) __a = AutoTokenizer.from_pretrained("google/umt5-small" , use_fast=lowerCamelCase , legacy=lowerCamelCase ) __a = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] __a = tokenizer(lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ).input_ids # fmt: off __a = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase , lowerCamelCase ) __a = model.generate(input_ids.to(lowerCamelCase ) ) __a = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] __a = tokenizer.batch_decode(lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import argparse __lowerCamelCase:Optional[Any] = """docs/source/_static/js/custom.js""" def _lowerCamelCase( a ): with open(__snake_case , encoding="utf-8" , newline="\n" ) as f: __a = f.readlines() __a = 0 # First let's put the right version while not lines[index].startswith("const stableVersion =" ): index += 1 __a = F"const stableVersion = \"v{version}\"\n" # Then update the dictionary while not lines[index].startswith("const versionMapping = {" ): index += 1 # We go until the end while not lines[index].startswith("}" ): index += 1 # We add the new version at the end lines[index - 1] += F" \"v{version}\": \"v{version}\",\n" with open(__snake_case , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(__snake_case ) if __name__ == "__main__": __lowerCamelCase:str = argparse.ArgumentParser() parser.add_argument("""--version""", help="""Release version.""") __lowerCamelCase:int = parser.parse_args() update_custom_js(args.version)
707
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase( a , a , a ): # Initialise PyTorch model __a = MobileBertConfig.from_json_file(a ) print(F"Building PyTorch model from configuration: {config}" ) __a = MobileBertForPreTraining(a ) # Load weights from tf checkpoint __a = load_tf_weights_in_mobilebert(a , a , a ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) SCREAMING_SNAKE_CASE__:List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
67
0
import re from filelock import FileLock try: import nltk SCREAMING_SNAKE_CASE__:Any = True except (ImportError, ModuleNotFoundError): SCREAMING_SNAKE_CASE__:Optional[Any] = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def _lowerCamelCase( a ): re.sub("<n>" , "" , UpperCAmelCase__ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(UpperCAmelCase__ ) )
708
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__ ( snake_case_ ): def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __a = input_file.read() __a = regexp.search(lowerCamelCase ) return match def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __a = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __a = regexp.finditer(lowerCamelCase ) __a = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCamelCase ) ): raise AssertionError(F"open(...) must use utf-8 encoding in {dataset}" ) def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCamelCase ) ): raise AssertionError(F"print statement found in {dataset}. Use datasets.logger/logging instead." )
67
0
"""simple docstring""" from transformers import BertTokenizer, EncoderDecoderModel, SeqaSeqTrainer, SeqaSeqTrainingArguments from transformers.testing_utils import TestCasePlus, require_torch, slow from transformers.utils import is_datasets_available if is_datasets_available(): import datasets class snake_case__ ( a__ ): @slow @require_torch def a__ ( self ): __a = EncoderDecoderModel.from_encoder_decoder_pretrained("prajjwal1/bert-tiny" , "prajjwal1/bert-tiny" ) __a = BertTokenizer.from_pretrained("bert-base-uncased" ) __a = bertabert.config.encoder.vocab_size __a = tokenizer.sep_token_id __a = tokenizer.cls_token_id __a = 128 __a = datasets.load_dataset("cnn_dailymail" , "3.0.0" , split="train[:1%]" ) __a = datasets.load_dataset("cnn_dailymail" , "3.0.0" , split="validation[:1%]" ) __a = train_dataset.select(range(32 ) ) __a = val_dataset.select(range(16 ) ) __a = 4 def _map_to_encoder_decoder_inputs(lowerCamelCase ): # Tokenizer will automatically set [BOS] <text> [EOS] __a = tokenizer(batch["article"] , padding="max_length" , truncation=lowerCAmelCase__ , max_length=512 ) __a = tokenizer(batch["highlights"] , padding="max_length" , truncation=lowerCAmelCase__ , max_length=128 ) __a = inputs.input_ids __a = inputs.attention_mask __a = outputs.input_ids __a = outputs.input_ids.copy() __a = [ [-100 if token == tokenizer.pad_token_id else token for token in labels] for labels in batch["labels"] ] __a = outputs.attention_mask assert all(len(lowerCAmelCase__ ) == 512 for x in inputs.input_ids ) assert all(len(lowerCAmelCase__ ) == 128 for x in outputs.input_ids ) return batch def _compute_metrics(lowerCamelCase ): __a = pred.label_ids __a = pred.predictions # all unnecessary tokens are removed __a = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) __a = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) __a = sum([int(pred_str[i] == label_str[i] ) for i in range(len(lowerCAmelCase__ ) )] ) / len(lowerCAmelCase__ ) return {"accuracy": accuracy} # map train dataset __a = train_dataset.map( _map_to_encoder_decoder_inputs , batched=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , remove_columns=["article", "highlights"] , ) train_dataset.set_format( type="torch" , columns=["input_ids", "attention_mask", "decoder_input_ids", "decoder_attention_mask", "labels"] , ) # same for validation dataset __a = val_dataset.map( _map_to_encoder_decoder_inputs , batched=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , remove_columns=["article", "highlights"] , ) val_dataset.set_format( type="torch" , columns=["input_ids", "attention_mask", "decoder_input_ids", "decoder_attention_mask", "labels"] , ) __a = self.get_auto_remove_tmp_dir() __a = SeqaSeqTrainingArguments( output_dir=lowerCAmelCase__ , per_device_train_batch_size=lowerCAmelCase__ , per_device_eval_batch_size=lowerCAmelCase__ , predict_with_generate=lowerCAmelCase__ , evaluation_strategy="steps" , do_train=lowerCAmelCase__ , do_eval=lowerCAmelCase__ , warmup_steps=0 , eval_steps=2 , logging_steps=2 , ) # instantiate trainer __a = SeqaSeqTrainer( model=lowerCAmelCase__ , args=lowerCAmelCase__ , compute_metrics=_compute_metrics , train_dataset=lowerCAmelCase__ , eval_dataset=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , ) # start training trainer.train()
709
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
67
0
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class snake_case__ ( UpperCAmelCase__ ): _snake_case : Union[str, Any] = (UniPCMultistepScheduler,) _snake_case : Union[str, Any] = (("""num_inference_steps""", 25),) def a__ ( self , **lowerCamelCase ): __a = { "num_train_timesteps": 1000, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", "solver_order": 2, "solver_type": "bh2", } config.update(**__lowerCAmelCase ) return config def a__ ( self , lowerCamelCase=0 , **lowerCamelCase ): __a = dict(self.forward_default_kwargs ) __a = kwargs.pop("num_inference_steps" , __lowerCAmelCase ) __a = self.dummy_sample __a = 0.1 * sample __a = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: __a = self.get_scheduler_config(**__lowerCAmelCase ) __a = scheduler_class(**__lowerCAmelCase ) scheduler.set_timesteps(__lowerCAmelCase ) # copy over dummy past residuals __a = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__lowerCAmelCase ) __a = scheduler_class.from_pretrained(__lowerCAmelCase ) new_scheduler.set_timesteps(__lowerCAmelCase ) # copy over dummy past residuals __a = dummy_past_residuals[: new_scheduler.config.solver_order] __a , __a = sample, sample for t in range(__lowerCAmelCase , time_step + scheduler.config.solver_order + 1 ): __a = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ).prev_sample __a = new_scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def a__ ( self , lowerCamelCase=0 , **lowerCamelCase ): __a = dict(self.forward_default_kwargs ) __a = kwargs.pop("num_inference_steps" , __lowerCAmelCase ) __a = self.dummy_sample __a = 0.1 * sample __a = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: __a = self.get_scheduler_config() __a = scheduler_class(**__lowerCAmelCase ) scheduler.set_timesteps(__lowerCAmelCase ) # copy over dummy past residuals (must be after setting timesteps) __a = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__lowerCAmelCase ) __a = scheduler_class.from_pretrained(__lowerCAmelCase ) # copy over dummy past residuals new_scheduler.set_timesteps(__lowerCAmelCase ) # copy over dummy past residual (must be after setting timesteps) __a = dummy_past_residuals[: new_scheduler.config.solver_order] __a = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ).prev_sample __a = new_scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def a__ ( self , lowerCamelCase=None , **lowerCamelCase ): if scheduler is None: __a = self.scheduler_classes[0] __a = self.get_scheduler_config(**__lowerCAmelCase ) __a = scheduler_class(**__lowerCAmelCase ) __a = self.scheduler_classes[0] __a = self.get_scheduler_config(**__lowerCAmelCase ) __a = scheduler_class(**__lowerCAmelCase ) __a = 10 __a = self.dummy_model() __a = self.dummy_sample_deter scheduler.set_timesteps(__lowerCAmelCase ) for i, t in enumerate(scheduler.timesteps ): __a = model(__lowerCAmelCase , __lowerCAmelCase ) __a = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).prev_sample return sample def a__ ( self ): __a = dict(self.forward_default_kwargs ) __a = kwargs.pop("num_inference_steps" , __lowerCAmelCase ) for scheduler_class in self.scheduler_classes: __a = self.get_scheduler_config() __a = scheduler_class(**__lowerCAmelCase ) __a = self.dummy_sample __a = 0.1 * sample if num_inference_steps is not None and hasattr(__lowerCAmelCase , "set_timesteps" ): scheduler.set_timesteps(__lowerCAmelCase ) elif num_inference_steps is not None and not hasattr(__lowerCAmelCase , "set_timesteps" ): __a = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) __a = [residual + 0.2, residual + 0.15, residual + 0.10] __a = dummy_past_residuals[: scheduler.config.solver_order] __a = scheduler.timesteps[5] __a = scheduler.timesteps[6] __a = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ).prev_sample __a = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def a__ ( self ): __a = UniPCMultistepScheduler(**self.get_scheduler_config() ) __a = self.full_loop(scheduler=__lowerCAmelCase ) __a = torch.mean(torch.abs(__lowerCAmelCase ) ) assert abs(result_mean.item() - 0.2464 ) < 1E-3 __a = DPMSolverSinglestepScheduler.from_config(scheduler.config ) __a = DEISMultistepScheduler.from_config(scheduler.config ) __a = DPMSolverMultistepScheduler.from_config(scheduler.config ) __a = UniPCMultistepScheduler.from_config(scheduler.config ) __a = self.full_loop(scheduler=__lowerCAmelCase ) __a = torch.mean(torch.abs(__lowerCAmelCase ) ) assert abs(result_mean.item() - 0.2464 ) < 1E-3 def a__ ( self ): for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=__lowerCAmelCase ) def a__ ( self ): self.check_over_configs(thresholding=__lowerCAmelCase ) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=__lowerCAmelCase , prediction_type=__lowerCAmelCase , sample_max_value=__lowerCAmelCase , solver_order=__lowerCAmelCase , solver_type=__lowerCAmelCase , ) def a__ ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__lowerCAmelCase ) def a__ ( self ): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=__lowerCAmelCase , solver_type=__lowerCAmelCase , prediction_type=__lowerCAmelCase , ) __a = self.full_loop( solver_order=__lowerCAmelCase , solver_type=__lowerCAmelCase , prediction_type=__lowerCAmelCase , ) assert not torch.isnan(__lowerCAmelCase ).any(), "Samples have nan numbers" def a__ ( self ): self.check_over_configs(lower_order_final=__lowerCAmelCase ) self.check_over_configs(lower_order_final=__lowerCAmelCase ) def a__ ( self ): for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=__lowerCAmelCase , time_step=0 ) def a__ ( self ): __a = self.full_loop() __a = torch.mean(torch.abs(__lowerCAmelCase ) ) assert abs(result_mean.item() - 0.2464 ) < 1E-3 def a__ ( self ): __a = self.full_loop(prediction_type="v_prediction" ) __a = torch.mean(torch.abs(__lowerCAmelCase ) ) assert abs(result_mean.item() - 0.1014 ) < 1E-3 def a__ ( self ): __a = self.scheduler_classes[0] __a = self.get_scheduler_config(thresholding=__lowerCAmelCase , dynamic_thresholding_ratio=0 ) __a = scheduler_class(**__lowerCAmelCase ) __a = 10 __a = self.dummy_model() __a = self.dummy_sample_deter.half() scheduler.set_timesteps(__lowerCAmelCase ) for i, t in enumerate(scheduler.timesteps ): __a = model(__lowerCAmelCase , __lowerCAmelCase ) __a = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ).prev_sample assert sample.dtype == torch.floataa def a__ ( self , **lowerCamelCase ): for scheduler_class in self.scheduler_classes: __a = self.get_scheduler_config(**__lowerCAmelCase ) __a = scheduler_class(**__lowerCAmelCase ) scheduler.set_timesteps(scheduler.config.num_train_timesteps ) assert len(scheduler.timesteps.unique() ) == scheduler.num_inference_steps
710
"""simple docstring""" import heapq import sys import numpy as np SCREAMING_SNAKE_CASE__:Optional[int] = tuple[int, int] class snake_case__ : def __init__( self ): __a = [] __a = set() def a__ ( self ): if not self.empty(): return self.elements[0][0] else: return float("inf" ) def a__ ( self ): return len(self.elements ) == 0 def a__ ( self , lowerCamelCase , lowerCamelCase ): if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(lowerCamelCase ) else: # update # print("update", item) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def a__ ( self , lowerCamelCase ): if item in self.set: self.set.remove(lowerCamelCase ) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def a__ ( self ): return self.elements[0][1] def a__ ( self ): ((__a) , (__a)) = heapq.heappop(self.elements ) self.set.remove(lowerCamelCase ) return (priority, item) def _lowerCamelCase( a , a ): # euclidean distance __a = np.array(a ) __a = np.array(a ) return np.linalg.norm(a - b ) def _lowerCamelCase( a , a ): # integer division by time variable return consistent_heuristic(a , a ) // t def _lowerCamelCase( a , a ): # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _lowerCamelCase( a , a , a , a ): __a = g_function[start] + Wa * heuristics[i](a , a ) return ans def _lowerCamelCase( a , a , a ): __a = np.chararray((n, n) ) for i in range(a ): for j in range(a ): __a = "*" for i in range(a ): for j in range(a ): if (j, (n - 1) - i) in blocks: __a = "#" __a = "-" __a = back_pointer[goal] while x != start: ((__a) , (__a)) = x # print(x) __a = "-" __a = back_pointer[x] __a = "-" for i in range(a ): for j in range(a ): if (i, j) == (0, n - 1): print(grid[i][j] , end=" " ) print("<-- End position" , end=" " ) else: print(grid[i][j] , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) print("PATH TAKEN BY THE ALGORITHM IS:-" ) __a = back_pointer[goal] while x != start: print(a , end=" " ) __a = back_pointer[x] print(a ) sys.exit() def _lowerCamelCase( a ): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _lowerCamelCase( a , a , a , a , a , a , a , a , ): for itera in range(a ): open_list[itera].remove_element(a ) # print("s", s) # print("j", j) ((__a) , (__a)) = s __a = (x - 1, y) __a = (x + 1, y) __a = (x, y + 1) __a = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(a ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(a ) __a = -1 __a = float("inf" ) if valid(a ) and g_function[neighbours] > g_function[s] + 1: __a = g_function[s] + 1 __a = s if neighbours not in close_list_anchor: open_list[0].put(a , key(a , 0 , a , a ) ) if neighbours not in close_list_inad: for var in range(1 , a ): if key(a , a , a , a ) <= Wa * key( a , 0 , a , a ): open_list[j].put( a , key(a , a , a , a ) ) def _lowerCamelCase( ): __a = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(1_5 , 2_0 ): some_list.append((x, 1_7) ) for x in range(1_0 , 1_9 ): for y in range(1 , 1_5 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(1_2 , 1_9 ): some_list.append((x, y) ) for x in range(3 , 1_3 ): for y in range(1_6 , 1_9 ): some_list.append((x, y) ) return some_list SCREAMING_SNAKE_CASE__:Any = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} SCREAMING_SNAKE_CASE__:str = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] SCREAMING_SNAKE_CASE__:int = make_common_ground() SCREAMING_SNAKE_CASE__:List[str] = blocks_blk # hyper parameters SCREAMING_SNAKE_CASE__:str = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 20 SCREAMING_SNAKE_CASE__:Dict = 3 # one consistent and two other inconsistent # start and end destination SCREAMING_SNAKE_CASE__:Dict = (0, 0) SCREAMING_SNAKE_CASE__:Optional[Any] = (n - 1, n - 1) SCREAMING_SNAKE_CASE__:List[str] = 1 def _lowerCamelCase( a , a , a ): __a = {start: 0, goal: float("inf" )} __a = {start: -1, goal: -1} __a = [] __a = set() for i in range(a ): open_list.append(PriorityQueue() ) open_list[i].put(a , key(a , a , a , a ) ) __a = [] __a = [] while open_list[0].minkey() < float("inf" ): for i in range(1 , a ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a , __a = open_list[i].top_show() visited.add(a ) expand_state( a , a , a , a , a , a , a , a , ) close_list_inad.append(a ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a = open_list[0].top_show() visited.add(a ) expand_state( a , 0 , a , a , a , a , a , a , ) close_list_anchor.append(a ) print("No path found to goal" ) print() for i in range(n - 1 , -1 , -1 ): for j in range(a ): if (j, i) in blocks: print("#" , end=" " ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" " ) else: print("-" , end=" " ) else: print("*" , end=" " ) if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
67
0
"""simple docstring""" from __future__ import annotations def _lowerCamelCase( a , a = None , a = None ): if start is None: __a = 0 if end is None: __a = len(_snake_case ) - 1 if start >= end: return __a = (start + end) // 2 slowsort(_snake_case , _snake_case , _snake_case ) slowsort(_snake_case , mid + 1 , _snake_case ) if sequence[end] < sequence[mid]: __a , __a = sequence[mid], sequence[end] slowsort(_snake_case , _snake_case , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
711
"""simple docstring""" SCREAMING_SNAKE_CASE__:Any = """Alexander Joslin""" import operator as op from .stack import Stack def _lowerCamelCase( a ): __a = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} __a = Stack() __a = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a ) ) elif i in operators: # RULE 2 operator_stack.push(a ) elif i == ")": # RULE 4 __a = operator_stack.peek() operator_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operators[opr](a , a ) operand_stack.push(a ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
67
0
"""simple docstring""" import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration SCREAMING_SNAKE_CASE__:Optional[int] = pytest.mark.integration SCREAMING_SNAKE_CASE__:int = {'comet'} SCREAMING_SNAKE_CASE__:Optional[int] = importlib.util.find_spec("""fairseq""") is not None SCREAMING_SNAKE_CASE__:Optional[Any] = {'code_eval'} SCREAMING_SNAKE_CASE__:List[str] = os.name == 'nt' SCREAMING_SNAKE_CASE__:Optional[Any] = {'bertscore', 'frugalscore', 'perplexity'} SCREAMING_SNAKE_CASE__:Dict = importlib.util.find_spec("""transformers""") is not None def _lowerCamelCase( a ): @wraps(_lowercase ) def wrapper(self , a ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("\"test requires Fairseq\"" ) else: test_case(self , _lowercase ) return wrapper def _lowerCamelCase( a ): @wraps(_lowercase ) def wrapper(self , a ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("\"test requires transformers\"" ) else: test_case(self , _lowercase ) return wrapper def _lowerCamelCase( a ): @wraps(_lowercase ) def wrapper(self , a ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("\"test not supported on Windows\"" ) else: test_case(self , _lowercase ) return wrapper def _lowerCamelCase( ): __a = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("./metrics/*/" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( UpperCamelCase_, UpperCamelCase_, UpperCamelCase_ ) @local class snake_case__ ( parameterized.TestCase ): _snake_case : Optional[int] = {} _snake_case : Any = None @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning" ) @pytest.mark.filterwarnings("ignore:load_metric is deprecated:FutureWarning" ) def a__ ( self , lowerCamelCase ): __a = '''[...]''' __a = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , UpperCamelCase__ ) ).module_path ) __a = datasets.load.import_main_class(metric_module.__name__ , dataset=UpperCamelCase__ ) # check parameters __a = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(UpperCamelCase__ , metric_module.__name__ ): with self.use_local_metrics(): try: __a = doctest.testmod(UpperCamelCase__ , verbose=UpperCamelCase__ , raise_on_error=UpperCamelCase__ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def a__ ( self , lowerCamelCase ): __a = '''[...]''' __a = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , UpperCamelCase__ ) ).module_path ) # run doctest with self.use_local_metrics(): __a = doctest.testmod(UpperCamelCase__ , verbose=UpperCamelCase__ , raise_on_error=UpperCamelCase__ ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def a__ ( self , lowerCamelCase , lowerCamelCase ): if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](UpperCamelCase__ ): yield else: yield @contextmanager def a__ ( self ): def load_local_metric(lowerCamelCase , *lowerCamelCase , **lowerCamelCase ): return load_metric(os.path.join("metrics" , UpperCamelCase__ ) , *UpperCamelCase__ , **UpperCamelCase__ ) with patch("datasets.load_metric" ) as mock_load_metric: __a = load_local_metric yield @classmethod def a__ ( cls , lowerCamelCase ): def wrapper(lowerCamelCase ): __a = contextmanager(UpperCamelCase__ ) __a = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("bleurt" ) def _lowerCamelCase( a ): import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("sv" , "" , "" ) # handle pytest cli flags class snake_case__ ( UpperCamelCase_ ): def a__ ( self , lowerCamelCase ): assert len(input_dict["input_ids"] ) == 2 return np.array([1.03, 1.04] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("bleurt.score._create_predictor" ) as mock_create_predictor: __a = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("bertscore" ) def _lowerCamelCase( a ): import torch def bert_cos_score_idf(a , a , *a , **a ): return torch.tensor([[1.0, 1.0, 1.0]] * len(_lowercase ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("bert_score.scorer.get_model" ), patch( "bert_score.scorer.bert_cos_score_idf" ) as mock_bert_cos_score_idf: __a = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("comet" ) def _lowerCamelCase( a ): def load_from_checkpoint(a ): class snake_case__ : def a__ ( self , lowerCamelCase , *lowerCamelCase , **lowerCamelCase ): assert len(UpperCamelCase__ ) == 2 __a = [0.19, 0.92] return scores, sum(UpperCamelCase__ ) / len(UpperCamelCase__ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("comet.download_model" ) as mock_download_model: __a = None with patch("comet.load_from_checkpoint" ) as mock_load_from_checkpoint: __a = load_from_checkpoint yield def _lowerCamelCase( ): __a = load_metric(os.path.join("metrics" , "seqeval" ) ) __a = '''ERROR''' __a = F"Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}" with pytest.raises(_lowercase , match=re.escape(_lowercase ) ): metric.compute(predictions=[] , references=[] , scheme=_lowercase )
712
"""simple docstring""" from math import pi def _lowerCamelCase( a , a ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
67
0
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def _lowerCamelCase( a ): __a = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "decoder.output_projection.weight", "_float_tensor", "encoder.embed_positions._float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(__UpperCamelCase , __UpperCamelCase ) def _lowerCamelCase( a ): __a , __a = emb.weight.shape __a = nn.Linear(__UpperCamelCase , __UpperCamelCase , bias=__UpperCamelCase ) __a = emb.weight.data return lin_layer def _lowerCamelCase( a ): __a = torch.load(__UpperCamelCase , map_location="cpu" ) __a = mam_aaa["args"] or mam_aaa["cfg"]["model"] __a = mam_aaa["model"] remove_ignore_keys_(__UpperCamelCase ) __a = state_dict["encoder.embed_tokens.weight"].shape[0] __a = MaMaaaConfig( vocab_size=__UpperCamelCase , max_position_embeddings=1_0_2_4 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="relu" , ) __a = state_dict["decoder.embed_tokens.weight"] __a = MaMaaaForConditionalGeneration(__UpperCamelCase ) model.model.load_state_dict(__UpperCamelCase , strict=__UpperCamelCase ) __a = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Any = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") SCREAMING_SNAKE_CASE__:List[str] = parser.parse_args() SCREAMING_SNAKE_CASE__:Dict = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
713
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): _snake_case : Dict = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = IMAGENET_DEFAULT_MEAN , lowerCamelCase = IMAGENET_DEFAULT_STD , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __a = image_std if image_std is not None else IMAGENET_DEFAULT_STD def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __a = int((256 / 224) * size["shortest_edge"] ) __a = get_resize_output_image_size(lowerCamelCase , size=lowerCamelCase , default_to_square=lowerCamelCase ) __a = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}" ) return resize( lowerCamelCase , size=(size_dict["height"], size_dict["width"]) , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"Size dict must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(lowerCamelCase , lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(lowerCamelCase , lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
"""simple docstring""" def _lowerCamelCase( a ): if len(a ) <= 1: return lst __a = 1 while i < len(a ): if lst[i - 1] <= lst[i]: i += 1 else: __a = lst[i], lst[i - 1] i -= 1 if i == 0: __a = 1 return lst if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Union[str, Any] = input("""Enter numbers separated by a comma:\n""").strip() SCREAMING_SNAKE_CASE__:Dict = [int(item) for item in user_input.split(""",""")] print(gnome_sort(unsorted))
714
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=30 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=None , lowerCamelCase=2 , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def a__ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, pixel_values, labels def a__ ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTForMaskedImageModeling(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __a = 1 __a = ViTForMaskedImageModeling(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.type_sequence_label_size __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a = 1 __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, unittest.TestCase ): _snake_case : Any = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) _snake_case : List[Any] = ( {"""feature-extraction""": ViTModel, """image-classification""": ViTForImageClassification} if is_torch_available() else {} ) _snake_case : int = True _snake_case : int = False _snake_case : str = False _snake_case : Optional[Any] = False def a__ ( self ): __a = ViTModelTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def a__ ( self ): pass def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def a__ ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def _lowerCamelCase( ): __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def a__ ( self ): return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def a__ ( self ): __a = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(lowerCamelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(**lowerCamelCase ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __a = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def a__ ( self ): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. __a = ViTModel.from_pretrained("facebook/dino-vits8" ).to(lowerCamelCase ) __a = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(lowerCamelCase , interpolate_pos_encoding=lowerCamelCase ) # verify the logits __a = torch.Size((1, 3601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , lowerCamelCase ) __a = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def a__ ( self ): __a = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __a = model(lowerCamelCase )
67
0
"""simple docstring""" import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class snake_case__ ( lowercase__ ): def __init__( self , lowerCamelCase = "▁" , lowerCamelCase = True , lowerCamelCase = "<unk>" , lowerCamelCase = "</s>" , lowerCamelCase = "<pad>" , ): __a = { "pad": {"id": 0, "token": pad_token}, "eos": {"id": 1, "token": eos_token}, "unk": {"id": 2, "token": unk_token}, } __a = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): __a = token_dict["token"] __a = Tokenizer(Unigram() ) __a = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(" {2,}" ) , " " ), normalizers.Lowercase(), ] ) __a = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__lowercase , add_prefix_space=__lowercase ), pre_tokenizers.Digits(individual_digits=__lowercase ), pre_tokenizers.Punctuation(), ] ) __a = decoders.Metaspace(replacement=__lowercase , add_prefix_space=__lowercase ) __a = TemplateProcessing( single=F"$A {self.special_tokens['eos']['token']}" , special_tokens=[(self.special_tokens["eos"]["token"], self.special_tokens["eos"]["id"])] , ) __a = { "model": "SentencePieceUnigram", "replacement": replacement, "add_prefix_space": add_prefix_space, } super().__init__(__lowercase , __lowercase ) def a__ ( self , lowerCamelCase , lowerCamelCase = 8000 , lowerCamelCase = True , ): __a = trainers.UnigramTrainer( vocab_size=__lowercase , special_tokens=self.special_tokens_list , show_progress=__lowercase , ) if isinstance(__lowercase , __lowercase ): __a = [files] self._tokenizer.train(__lowercase , trainer=__lowercase ) self.add_unk_id() def a__ ( self , lowerCamelCase , lowerCamelCase = 8000 , lowerCamelCase = True , ): __a = trainers.UnigramTrainer( vocab_size=__lowercase , special_tokens=self.special_tokens_list , show_progress=__lowercase , ) self._tokenizer.train_from_iterator(__lowercase , trainer=__lowercase ) self.add_unk_id() def a__ ( self ): __a = json.loads(self._tokenizer.to_str() ) __a = self.special_tokens["unk"]["id"] __a = Tokenizer.from_str(json.dumps(__lowercase ) )
715
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=True , lowerCamelCase=1 / 255 , lowerCamelCase=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __a = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __a = parent __a = batch_size __a = num_channels __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std __a = do_rescale __a = rescale_factor __a = do_pad def a__ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def a__ ( self , lowerCamelCase , lowerCamelCase=False ): if not batched: __a = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __a , __a = image.size else: __a , __a = image.shape[1], image.shape[2] if w < h: __a = int(self.size["shortest_edge"] * h / w ) __a = self.size["shortest_edge"] elif w > h: __a = self.size["shortest_edge"] __a = int(self.size["shortest_edge"] * w / h ) else: __a = self.size["shortest_edge"] __a = self.size["shortest_edge"] else: __a = [] for image in image_inputs: __a , __a = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : List[Any] = DetaImageProcessor if is_vision_available() else None def a__ ( self ): __a = DetaImageProcessingTester(self ) @property def a__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_rescale" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_pad" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def a__ ( self ): __a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def a__ ( self ): pass def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self ): # prepare image and target __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"image_id": 39769, "annotations": target} # encode them __a = DetaImageProcessor() __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def a__ ( self ): # prepare image, target and masks_path __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __a = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __a = DetaImageProcessor(format="coco_panoptic" ) __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __a = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
67
0
"""simple docstring""" import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class UpperCamelCase ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=18 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=[0.5, 0.5, 0.5] , ): __a = size if size is not None else {"height": 18, "width": 18} __a = parent __a = batch_size __a = num_channels __a = image_size __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std def a__ ( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class UpperCamelCase ( __snake_case, unittest.TestCase ): _snake_case : Optional[Any] = DPTImageProcessor if is_vision_available() else None def a__ ( self ): __a = DPTImageProcessingTester(self ) @property def a__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(A_ , "image_mean" ) ) self.assertTrue(hasattr(A_ , "image_std" ) ) self.assertTrue(hasattr(A_ , "do_normalize" ) ) self.assertTrue(hasattr(A_ , "do_resize" ) ) self.assertTrue(hasattr(A_ , "size" ) ) def a__ ( self ): __a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) __a = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ ) for image in image_inputs: self.assertIsInstance(A_ , Image.Image ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __a = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , numpify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , np.ndarray ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __a = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=A_ , torchify=A_ ) for image in image_inputs: self.assertIsInstance(A_ , torch.Tensor ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __a = image_processing(A_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
716
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE__:Dict = logging.getLogger() def _lowerCamelCase( ): __a = argparse.ArgumentParser() parser.add_argument("-f" ) __a = parser.parse_args() return args.f class snake_case__ ( snake_case_ ): def a__ ( self ): __a = logging.StreamHandler(sys.stdout ) logger.addHandler(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowerCamelCase , "argv" , lowerCamelCase ): __a = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self ): __a = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase )
67
0
import warnings from functools import wraps from typing import Callable def _lowerCamelCase( a ): @wraps(a ) def _inner_fn(*a , **a ): warnings.warn( (F"\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.") , a , ) return fn(*a , **a ) return _inner_fn
717
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __a = image_std if image_std is not None else OPENAI_CLIP_STD __a = do_convert_rgb def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) __a = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , param_name="size" , default_to_square=lowerCamelCase ) __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" , default_to_square=lowerCamelCase ) __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __a = [convert_to_rgb(lowerCamelCase ) for image in images] # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__:Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:List[str] = {"""vocab_file""": """vocab.json"""} SCREAMING_SNAKE_CASE__:Union[str, Any] = { """vocab_file""": { """mgp-str""": """https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json""", } } SCREAMING_SNAKE_CASE__:List[str] = {"""mgp-str""": 27} class snake_case__ ( UpperCAmelCase__ ): _snake_case : Any = VOCAB_FILES_NAMES _snake_case : List[str] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowerCamelCase , lowerCamelCase="[GO]" , lowerCamelCase="[GO]" , lowerCamelCase="[s]" , lowerCamelCase="[GO]" , **lowerCamelCase ): super().__init__( unk_token=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding="utf-8" ) as vocab_handle: __a = json.load(lowerCamelCase__ ) __a = {v: k for k, v in self.vocab.items()} @property def a__ ( self ): return len(self.vocab ) def a__ ( self ): return dict(self.vocab , **self.added_tokens_encoder ) def a__ ( self , lowerCamelCase ): __a = [] for s in text: char_tokens.extend(lowerCamelCase__ ) return char_tokens def a__ ( self , lowerCamelCase ): return self.vocab.get(lowerCamelCase__ , self.vocab.get(self.unk_token ) ) def a__ ( self , lowerCamelCase ): return self.decoder.get(lowerCamelCase__ ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error("Vocabulary path ({}) should be a directory".format(lowerCamelCase__ ) ) return __a = os.path.join( lowerCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowerCamelCase__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + "\n" ) return (vocab_file,)
718
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Any = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Optional[Any] = { """vocab_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/vocab.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/vocab.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/vocab.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/vocab.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/vocab.json""", }, """merges_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/merges.txt""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/merges.txt""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/merges.txt""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/merges.txt""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/merges.txt""", }, """tokenizer_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/tokenizer.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/tokenizer.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__:Union[str, Any] = { """gpt2""": 1024, """gpt2-medium""": 1024, """gpt2-large""": 1024, """gpt2-xl""": 1024, """distilgpt2""": 1024, } class snake_case__ ( snake_case_ ): _snake_case : Tuple = VOCAB_FILES_NAMES _snake_case : str = PRETRAINED_VOCAB_FILES_MAP _snake_case : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : List[str] = ["""input_ids""", """attention_mask"""] _snake_case : Dict = GPTaTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , add_prefix_space=lowerCamelCase , **lowerCamelCase , ) __a = kwargs.pop("add_bos_token" , lowerCamelCase ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__:Any = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:int = """https://openaipublic.azureedge.net/jukebox/models/""" SCREAMING_SNAKE_CASE__:List[str] = { """jukebox-1b-lyrics""": [ """5b/vqvae.pth.tar""", """5b/prior_level_0.pth.tar""", """5b/prior_level_1.pth.tar""", """1b_lyrics/prior_level_2.pth.tar""", ], """jukebox-5b-lyrics""": [ """5b/vqvae.pth.tar""", """5b/prior_level_0.pth.tar""", """5b/prior_level_1.pth.tar""", """5b_lyrics/prior_level_2.pth.tar""", ], } def _lowerCamelCase( a ): if key.endswith(".model.1.bias" ) and len(key.split("." ) ) > 1_0: __a = key.replace(".model.1.bias" , ".conv1d_1.bias" ) elif key.endswith(".model.1.weight" ) and len(key.split("." ) ) > 1_0: __a = key.replace(".model.1.weight" , ".conv1d_1.weight" ) elif key.endswith(".model.3.bias" ) and len(key.split("." ) ) > 1_0: __a = key.replace(".model.3.bias" , ".conv1d_2.bias" ) elif key.endswith(".model.3.weight" ) and len(key.split("." ) ) > 1_0: __a = key.replace(".model.3.weight" , ".conv1d_2.weight" ) if "conditioner_blocks.0." in key: __a = key.replace("conditioner_blocks.0" , "conditioner_blocks" ) if "prime_prior" in key: __a = key.replace("prime_prior" , "encoder" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: __a = key.replace(".emb." , "." ) if key.endswith("k" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(".k" , ".codebook" ) if "y_emb." in key: return key.replace("y_emb." , "metadata_embedding." ) if "x_emb.emb." in key: __a = key.replace("0.x_emb.emb" , "embed_tokens" ) if "prime_state_ln" in key: return key.replace("prime_state_ln" , "encoder.final_layer_norm" ) if ".ln" in key: return key.replace(".ln" , ".layer_norm" ) if "_ln" in key: return key.replace("_ln" , "_layer_norm" ) if "prime_state_proj" in key: return key.replace("prime_state_proj" , "encoder.proj_in" ) if "prime_x_out" in key: return key.replace("prime_x_out" , "encoder.lm_head" ) if "prior.x_out" in key: return key.replace("x_out" , "fc_proj_out" ) if "x_emb" in key: return key.replace("x_emb" , "embed_tokens" ) return key def _lowerCamelCase( a , a , a , a ): __a = {} import re __a = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) __a = re.compile( R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) __a = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) __a = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) __a = re.compile( R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) __a = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) __a = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)" ) __a = re.compile( R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) __a = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(lowercase__ ): __a = re_encoder_block_conv_in.match(lowercase__ ) __a = regex_match.groups() __a = int(groups[2] ) * 2 + int(groups[3] ) __a = F"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}" __a = re_encoder_block_conv_in.sub(lowercase__ , lowercase__ ) elif re_encoder_block_resnet.fullmatch(lowercase__ ): __a = re_encoder_block_resnet.match(lowercase__ ) __a = regex_match.groups() __a = int(groups[2] ) * 2 + int(groups[3] ) __a = {"1": 1, "3": 2}[groups[-2]] __a = F"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}." __a = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __a = prefix + resnet_block __a = re_encoder_block_resnet.sub(lowercase__ , lowercase__ ) elif re_encoder_block_proj_out.fullmatch(lowercase__ ): __a = re_encoder_block_proj_out.match(lowercase__ ) __a = regex_match.groups() __a = F"encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}" __a = re_encoder_block_proj_out.sub(lowercase__ , lowercase__ ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(lowercase__ ): __a = re_decoder_block_conv_out.match(lowercase__ ) __a = regex_match.groups() __a = int(groups[2] ) * 2 + int(groups[3] ) - 2 __a = F"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}" __a = re_decoder_block_conv_out.sub(lowercase__ , lowercase__ ) elif re_decoder_block_resnet.fullmatch(lowercase__ ): __a = re_decoder_block_resnet.match(lowercase__ ) __a = regex_match.groups() __a = int(groups[2] ) * 2 + int(groups[3] ) - 2 __a = {"1": 1, "3": 2}[groups[-2]] __a = F"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}." __a = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __a = prefix + resnet_block __a = re_decoder_block_resnet.sub(lowercase__ , lowercase__ ) elif re_decoder_block_proj_in.fullmatch(lowercase__ ): __a = re_decoder_block_proj_in.match(lowercase__ ) __a = regex_match.groups() __a = F"decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}" __a = re_decoder_block_proj_in.sub(lowercase__ , lowercase__ ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(lowercase__ ): __a = re_prior_cond_conv_out.match(lowercase__ ) __a = regex_match.groups() __a = int(groups[1] ) * 2 + int(groups[2] ) - 2 __a = F"conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}" __a = re_prior_cond_conv_out.sub(lowercase__ , lowercase__ ) elif re_prior_cond_resnet.fullmatch(lowercase__ ): __a = re_prior_cond_resnet.match(lowercase__ ) __a = regex_match.groups() __a = int(groups[1] ) * 2 + int(groups[2] ) - 2 __a = {"1": 1, "3": 2}[groups[-2]] __a = F"conditioner_blocks.upsampler.upsample_block.{block_index}." __a = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __a = prefix + resnet_block __a = re_prior_cond_resnet.sub(lowercase__ , lowercase__ ) elif re_prior_cond_proj_in.fullmatch(lowercase__ ): __a = re_prior_cond_proj_in.match(lowercase__ ) __a = regex_match.groups() __a = F"conditioner_blocks.upsampler.proj_in.{groups[-1]}" __a = re_prior_cond_proj_in.sub(lowercase__ , lowercase__ ) # keep original key else: __a = original_key __a = replace_key(lowercase__ ) if F"{key_prefix}.{key}" not in model_state_dict or key is None: print(F"failed converting {original_key} to {key}, does not match" ) # handle missmatched shape elif value.shape != model_state_dict[F"{key_prefix}.{key}"].shape: __a = model_state_dict[F"{key_prefix}.{key}"] print(F"{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match" ) __a = original_key __a = original_key __a = value return new_dict @torch.no_grad() def _lowerCamelCase( a=None , a=None ): for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" ): __a = requests.get(F"{PREFIX}{file}" , allow_redirects=lowercase__ ) os.makedirs(F"{pytorch_dump_folder_path}/" , exist_ok=lowercase__ ) open(F"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" , "wb" ).write(r.content ) __a = MODEL_MAPPING[model_name.split("/" )[-1]] __a = JukeboxConfig.from_pretrained(lowercase__ ) __a = JukeboxModel(lowercase__ ) __a = [] __a = {} for i, dict_name in enumerate(lowercase__ ): __a = torch.load(F"{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}" )["model"] __a = {} for k in old_dic.keys(): if k.endswith(".b" ): __a = old_dic[k] elif k.endswith(".w" ): __a = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: __a = old_dic[k] else: __a = old_dic[k] __a = "vqvae" if i == 0 else F"priors.{3 - i}" __a = fix_jukebox_keys(lowercase__ , model.state_dict() , lowercase__ , lowercase__ ) weight_dict.append(lowercase__ ) __a = weight_dict.pop(0 ) model.vqvae.load_state_dict(lowercase__ ) for i in range(len(lowercase__ ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) with open(F"{pytorch_dump_folder_path}/mapping.json" , "w" ) as txtfile: json.dump(lowercase__ , lowercase__ ) print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(lowercase__ ) return weight_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""jukebox-5b-lyrics""", type=str, help="""Name of the model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""jukebox-5b-lyrics-converted""", type=str, help="""Path to the output PyTorch model directory.""", ) SCREAMING_SNAKE_CASE__:Dict = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
719
"""simple docstring""" from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("repo_id" , ["canonical_dataset_name", "org-name/dataset-name"] ) @pytest.mark.parametrize("path" , ["filename.csv", "filename with blanks.csv"] ) @pytest.mark.parametrize("revision" , [None, "v2"] ) def _lowerCamelCase( a , a , a ): __a = hf_hub_url(repo_id=a , path=a , revision=a ) assert url == F"https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(a )}"
67
0
"""simple docstring""" from PIL import Image def _lowerCamelCase( a , a ): def brightness(a ) -> float: return 1_2_8 + level + (c - 1_2_8) if not -2_5_5.0 <= level <= 2_5_5.0: raise ValueError("level must be between -255.0 (black) and 255.0 (white)" ) return img.point(UpperCamelCase__ ) if __name__ == "__main__": # Load image with Image.open("""image_data/lena.jpg""") as img: # Change brightness to 100 SCREAMING_SNAKE_CASE__:Optional[Any] = change_brightness(img, 100) brigt_img.save("""image_data/lena_brightness.png""", format="""png""")
720
"""simple docstring""" from __future__ import annotations def _lowerCamelCase( a , a , a ): if len(a ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(a ) or left < -len(a ) or right >= len(a ) or right < -len(a ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] __a = (left + right) >> 1 # the middle __a = find_max(a , a , a ) # find max in range[left, mid] __a = find_max(a , mid + 1 , a ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
67
0
"""simple docstring""" import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class snake_case__ ( unittest.TestCase ): def a__ ( self ): __a = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __a = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(UpperCAmelCase__ ) __a = -1 __a = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(UpperCAmelCase__ ) __a = model.generate(UpperCAmelCase__ , max_new_tokens=10 , do_sample=UpperCAmelCase__ ) __a = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: __a = TextStreamer(UpperCAmelCase__ ) model.generate(UpperCAmelCase__ , max_new_tokens=10 , do_sample=UpperCAmelCase__ , streamer=UpperCAmelCase__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __a = cs.out[:-1] self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __a = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(UpperCAmelCase__ ) __a = -1 __a = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(UpperCAmelCase__ ) __a = model.generate(UpperCAmelCase__ , max_new_tokens=10 , do_sample=UpperCAmelCase__ ) __a = tokenizer.decode(greedy_ids[0] ) __a = TextIteratorStreamer(UpperCAmelCase__ ) __a = {"input_ids": input_ids, "max_new_tokens": 10, "do_sample": False, "streamer": streamer} __a = Thread(target=model.generate , kwargs=UpperCAmelCase__ ) thread.start() __a = "" for new_text in streamer: streamer_text += new_text self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __a = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(UpperCAmelCase__ ) __a = -1 __a = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(UpperCAmelCase__ ) __a = model.generate(UpperCAmelCase__ , max_new_tokens=10 , do_sample=UpperCAmelCase__ ) __a = greedy_ids[:, input_ids.shape[1] :] __a = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: __a = TextStreamer(UpperCAmelCase__ , skip_prompt=UpperCAmelCase__ ) model.generate(UpperCAmelCase__ , max_new_tokens=10 , do_sample=UpperCAmelCase__ , streamer=UpperCAmelCase__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __a = cs.out[:-1] self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them __a = AutoTokenizer.from_pretrained("distilgpt2" ) __a = AutoModelForCausalLM.from_pretrained("distilgpt2" ).to(UpperCAmelCase__ ) __a = -1 __a = torch.ones((1, 5) , device=UpperCAmelCase__ ).long() * model.config.bos_token_id with CaptureStdout() as cs: __a = TextStreamer(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ ) model.generate(UpperCAmelCase__ , max_new_tokens=1 , do_sample=UpperCAmelCase__ , streamer=UpperCAmelCase__ ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token __a = cs.out[:-1] # Remove the final "\n" __a = tokenizer(UpperCAmelCase__ , return_tensors="pt" ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def a__ ( self ): __a = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __a = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(UpperCAmelCase__ ) __a = -1 __a = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(UpperCAmelCase__ ) __a = TextIteratorStreamer(UpperCAmelCase__ , timeout=0.001 ) __a = {"input_ids": input_ids, "max_new_tokens": 10, "do_sample": False, "streamer": streamer} __a = Thread(target=model.generate , kwargs=UpperCAmelCase__ ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(UpperCAmelCase__ ): __a = "" for new_text in streamer: streamer_text += new_text
721
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json""", """google/bigbird-roberta-large""": """https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json""", """google/bigbird-base-trivia-itc""": """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json""", # See all BigBird models at https://huggingface.co/models?filter=big_bird } class snake_case__ ( snake_case_ ): _snake_case : Any = """big_bird""" def __init__( self , lowerCamelCase=50358 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu_new" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=4096 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=True , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=66 , lowerCamelCase="block_sparse" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=64 , lowerCamelCase=3 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , sep_token_id=lowerCamelCase , **lowerCamelCase , ) __a = vocab_size __a = max_position_embeddings __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = initializer_range __a = type_vocab_size __a = layer_norm_eps __a = use_cache __a = rescale_embeddings __a = attention_type __a = use_bias __a = block_size __a = num_random_blocks __a = classifier_dropout class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task == "multiple-choice": __a = {0: "batch", 1: "choice", 2: "sequence"} else: __a = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
67
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( ImageTextPipelineOutput, UniDiffuserPipeline, ) else: from .modeling_text_decoder import UniDiffuserTextDecoder from .modeling_uvit import UniDiffuserModel, UTransformeraDModel from .pipeline_unidiffuser import ImageTextPipelineOutput, UniDiffuserPipeline
700
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:Optional[int] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[int] = {"""tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Tuple = { """tokenizer_file""": { """bigscience/tokenizer""": """https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json""", """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json""", }, } class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = VOCAB_FILES_NAMES _snake_case : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Optional[int] = ["""input_ids""", """attention_mask"""] _snake_case : Optional[int] = None def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<unk>" , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="<pad>" , lowerCamelCase=False , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , pad_token=lowerCamelCase , add_prefix_space=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , **lowerCamelCase , ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
"""simple docstring""" import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE__:Dict = """▁""" SCREAMING_SNAKE_CASE__:Union[str, Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") @require_sentencepiece @require_tokenizers class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : str = BigBirdTokenizer _snake_case : Optional[Any] = BigBirdTokenizerFast _snake_case : List[str] = True _snake_case : Tuple = True def a__ ( self ): super().setUp() __a = self.tokenizer_class(lowerCamelCase , keep_accents=lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def a__ ( self ): __a = "<s>" __a = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase ) , lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase ) , lowerCamelCase ) def a__ ( self ): __a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(lowerCamelCase ) , 1004 ) def a__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def a__ ( self ): if not self.test_rust_tokenizer: return __a = self.get_tokenizer() __a = self.get_rust_tokenizer() __a = "I was born in 92000, and this is falsé." __a = tokenizer.tokenize(lowerCamelCase ) __a = rust_tokenizer.tokenize(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __a = tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __a = rust_tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __a = self.get_rust_tokenizer() __a = tokenizer.encode(lowerCamelCase ) __a = rust_tokenizer.encode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = BigBirdTokenizer(lowerCamelCase , keep_accents=lowerCamelCase ) __a = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase ) , [285, 46, 10, 170, 382] , ) __a = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __a = tokenizer.convert_tokens_to_ids(lowerCamelCase ) self.assertListEqual( lowerCamelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) __a = tokenizer.convert_ids_to_tokens(lowerCamelCase ) self.assertListEqual( lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def a__ ( self ): return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def a__ ( self ): __a = "Hello World!" __a = [65, 18536, 2260, 101, 66] self.assertListEqual(lowerCamelCase , self.big_tokenizer.encode(lowerCamelCase ) ) @slow def a__ ( self ): __a = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off __a = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(lowerCamelCase , self.big_tokenizer.encode(lowerCamelCase ) ) @require_torch @slow def a__ ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence __a = list(self.big_tokenizer.get_vocab().keys() )[:10] __a = " ".join(lowerCamelCase ) __a = self.big_tokenizer.encode_plus(lowerCamelCase , return_tensors="pt" , return_token_type_ids=lowerCamelCase ) __a = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=lowerCamelCase ) __a = BigBirdConfig(attention_type="original_full" ) __a = BigBirdModel(lowerCamelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowerCamelCase ) model(**lowerCamelCase ) @slow def a__ ( self ): __a = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) __a = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def a__ ( self ): __a = {"input_ids": [[65, 39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114, 66], [65, 448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
701
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class snake_case__ : _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : int _snake_case : int _snake_case : float _snake_case : float _snake_case : Tuple[int] def a__ ( self ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def a__ ( self ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def a__ ( self ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def a__ ( self ): __a = torch.arange(self.height * self.width ) __a = torch.stack( [ pixel_indices % self.width, torch.div(lowerCamelCase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def a__ ( self ): __a , *__a = self.shape __a = int(np.prod(lowerCamelCase ) ) __a = self.get_image_coords() __a = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __a = self.get_camera_rays(lowerCamelCase ) __a = rays.view(lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def a__ ( self , lowerCamelCase ): __a , *__a , __a = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __a = coords.view(lowerCamelCase , -1 , 2 ) __a = self.resolution() __a = self.fov() __a = (flat.float() / (res - 1)) * 2 - 1 __a = fracs * torch.tan(fov / 2 ) __a = fracs.view(lowerCamelCase , -1 , 2 ) __a = ( self.z.view(lowerCamelCase , 1 , 3 ) + self.x.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) __a = directions / directions.norm(dim=-1 , keepdim=lowerCamelCase ) __a = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCamelCase , *lowerCamelCase , 2 , 3 ) def a__ ( self , lowerCamelCase , lowerCamelCase ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCamelCase , height=lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowerCamelCase( a ): __a = [] __a = [] __a = [] __a = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __a = np.array([np.sin(a ), np.cos(a ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __a = -z * 4 __a = np.array([np.cos(a ), -np.sin(a ), 0.0] ) __a = np.cross(a , a ) origins.append(a ) xs.append(a ) ys.append(a ) zs.append(a ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(a , axis=0 ) ).float() , x=torch.from_numpy(np.stack(a , axis=0 ) ).float() , y=torch.from_numpy(np.stack(a , axis=0 ) ).float() , z=torch.from_numpy(np.stack(a , axis=0 ) ).float() , width=a , height=a , x_fov=0.7 , y_fov=0.7 , shape=(1, len(a )) , )
67
0
"""simple docstring""" from math import log from scipy.constants import Boltzmann, physical_constants SCREAMING_SNAKE_CASE__:Union[str, Any] = 300 # TEMPERATURE (unit = K) def _lowerCamelCase( a , a , a , ): if donor_conc <= 0: raise ValueError("Donor concentration should be positive" ) elif acceptor_conc <= 0: raise ValueError("Acceptor concentration should be positive" ) elif intrinsic_conc <= 0: raise ValueError("Intrinsic concentration should be positive" ) elif donor_conc <= intrinsic_conc: raise ValueError( "Donor concentration should be greater than intrinsic concentration" ) elif acceptor_conc <= intrinsic_conc: raise ValueError( "Acceptor concentration should be greater than intrinsic concentration" ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
702
"""simple docstring""" def _lowerCamelCase( a ): return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase( a ): __a = 0 __a = number while duplicate > 0: __a , __a = divmod(a , 1_0 ) fact_sum += factorial(a ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__:Optional[Any] = int(input("""Enter number: """).strip()) print( F'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
67
0
"""simple docstring""" import math import tensorflow as tf from packaging import version def _lowerCamelCase( a ): __a = tf.convert_to_tensor(lowerCAmelCase_ ) __a = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def _lowerCamelCase( a ): __a = tf.convert_to_tensor(lowerCAmelCase_ ) __a = tf.cast(math.pi , x.dtype ) __a = tf.cast(0.04_47_15 , x.dtype ) __a = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(lowerCAmelCase_ , 3 )) )) return x * cdf def _lowerCamelCase( a ): __a = tf.convert_to_tensor(lowerCAmelCase_ ) return x * tf.tanh(tf.math.softplus(lowerCAmelCase_ ) ) def _lowerCamelCase( a ): __a = tf.convert_to_tensor(lowerCAmelCase_ ) __a = tf.cast(0.04_47_15 , x.dtype ) __a = tf.cast(0.79_78_84_56_08 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def _lowerCamelCase( a ): __a = tf.convert_to_tensor(lowerCAmelCase_ ) __a = tf.cast(1.7_02 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def _lowerCamelCase( a ): return tf.clip_by_value(_gelu(lowerCAmelCase_ ) , -1_0 , 1_0 ) def _lowerCamelCase( a , a=-1 ): __a = tf.split(lowerCAmelCase_ , 2 , axis=lowerCAmelCase_ ) return a * tf.math.sigmoid(lowerCAmelCase_ ) if version.parse(tf.version.VERSION) >= version.parse("""2.4"""): def _lowerCamelCase( a ): return tf.keras.activations.gelu(lowerCAmelCase_ , approximate=lowerCAmelCase_ ) SCREAMING_SNAKE_CASE__:int = tf.keras.activations.gelu SCREAMING_SNAKE_CASE__:Any = approximate_gelu_wrap else: SCREAMING_SNAKE_CASE__:List[Any] = _gelu SCREAMING_SNAKE_CASE__:Any = _gelu_new SCREAMING_SNAKE_CASE__:int = { 'gelu': gelu, 'gelu_10': gelu_aa, 'gelu_fast': gelu_fast, 'gelu_new': gelu_new, 'glu': glu, 'mish': mish, 'quick_gelu': quick_gelu, 'relu': tf.keras.activations.relu, 'sigmoid': tf.keras.activations.sigmoid, 'silu': tf.keras.activations.swish, 'swish': tf.keras.activations.swish, 'tanh': tf.keras.activations.tanh, } def _lowerCamelCase( a ): if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F"function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}" )
703
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE__:Optional[Any] = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Union[str, Any] = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" def _lowerCamelCase( a ): __a = len(SCREAMING_SNAKE_CASE__ ) for i in range(SCREAMING_SNAKE_CASE__ ): for j in range(i + 1 , SCREAMING_SNAKE_CASE__ ): if numbers[j] < numbers[i]: __a = numbers[j], numbers[i] return numbers if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = input("""Enter numbers separated by a comma:\n""").strip() SCREAMING_SNAKE_CASE__:Any = [int(item) for item in user_input.split(""",""")] print(exchange_sort(unsorted))
704
"""simple docstring""" import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def _lowerCamelCase( a , a , a ): __a = OmegaConf.load(a ) __a = torch.load(a , map_location="cpu" )["model"] __a = list(state_dict.keys() ) # extract state_dict for VQVAE __a = {} __a = "first_stage_model." for key in keys: if key.startswith(a ): __a = state_dict[key] # extract state_dict for UNetLDM __a = {} __a = "model.diffusion_model." for key in keys: if key.startswith(a ): __a = state_dict[key] __a = config.model.params.first_stage_config.params __a = config.model.params.unet_config.params __a = VQModel(**a ).eval() vqvae.load_state_dict(a ) __a = UNetLDMModel(**a ).eval() unet.load_state_dict(a ) __a = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=a , ) __a = LDMPipeline(a , a , a ) pipeline.save_pretrained(a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
67
0
"""simple docstring""" import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def _lowerCamelCase( a , a , a ): if isinstance(a_ , torch.Tensor ): return image elif isinstance(a_ , PIL.Image.Image ): __a = [image] if isinstance(image[0] , PIL.Image.Image ): __a = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) )[None, :] for i in image] __a = np.concatenate(a_ , axis=0 ) __a = np.array(a_ ).astype(np.floataa ) / 2_55.0 __a = image.transpose(0 , 3 , 1 , 2 ) __a = 2.0 * image - 1.0 __a = torch.from_numpy(a_ ) elif isinstance(image[0] , torch.Tensor ): __a = torch.cat(a_ , dim=0 ) return image def _lowerCamelCase( a , a , a , a=0.99_95 ): if not isinstance(a_ , np.ndarray ): __a = True __a = va.device __a = va.cpu().numpy() __a = va.cpu().numpy() __a = np.sum(va * va / (np.linalg.norm(a_ ) * np.linalg.norm(a_ )) ) if np.abs(a_ ) > DOT_THRESHOLD: __a = (1 - t) * va + t * va else: __a = np.arccos(a_ ) __a = np.sin(a_ ) __a = theta_a * t __a = np.sin(a_ ) __a = np.sin(theta_a - theta_t ) / sin_theta_a __a = sin_theta_t / sin_theta_a __a = sa * va + sa * va if inputs_are_torch: __a = torch.from_numpy(a_ ).to(a_ ) return va def _lowerCamelCase( a , a ): __a = F.normalize(a_ , dim=-1 ) __a = F.normalize(a_ , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def _lowerCamelCase( a , a ): for param in model.parameters(): __a = value class snake_case__ ( __SCREAMING_SNAKE_CASE ): def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): super().__init__() self.register_modules( vae=__snake_case , text_encoder=__snake_case , clip_model=__snake_case , tokenizer=__snake_case , unet=__snake_case , scheduler=__snake_case , feature_extractor=__snake_case , coca_model=__snake_case , coca_tokenizer=__snake_case , coca_transform=__snake_case , ) __a = ( feature_extractor.size if isinstance(feature_extractor.size , __snake_case ) else feature_extractor.size['''shortest_edge'''] ) __a = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , __snake_case ) set_requires_grad(self.clip_model , __snake_case ) def a__ ( self , lowerCamelCase = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __a = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__snake_case ) def a__ ( self ): self.enable_attention_slicing(__snake_case ) def a__ ( self ): set_requires_grad(self.vae , __snake_case ) def a__ ( self ): set_requires_grad(self.vae , __snake_case ) def a__ ( self ): set_requires_grad(self.unet , __snake_case ) def a__ ( self ): set_requires_grad(self.unet , __snake_case ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): # get the original timestep using init_timestep __a = min(int(num_inference_steps * strength ) , __snake_case ) __a = max(num_inference_steps - init_timestep , 0 ) __a = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None ): if not isinstance(__snake_case , torch.Tensor ): raise ValueError(F"`image` has to be of type `torch.Tensor` but is {type(__snake_case )}" ) __a = image.to(device=__snake_case , dtype=__snake_case ) if isinstance(__snake_case , __snake_case ): __a = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(__snake_case ) ] __a = torch.cat(__snake_case , dim=0 ) else: __a = self.vae.encode(__snake_case ).latent_dist.sample(__snake_case ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __a = 0.1_8215 * init_latents __a = init_latents.repeat_interleave(__snake_case , dim=0 ) __a = randn_tensor(init_latents.shape , generator=__snake_case , device=__snake_case , dtype=__snake_case ) # get latents __a = self.scheduler.add_noise(__snake_case , __snake_case , __snake_case ) __a = init_latents return latents def a__ ( self , lowerCamelCase ): __a = self.coca_transform(__snake_case ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): __a = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) __a = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("<end_of_text>" )[0].replace("<start_of_text>" , "" ).rstrip(" .," ) def a__ ( self , lowerCamelCase , lowerCamelCase ): __a = self.feature_extractor.preprocess(__snake_case ) __a = torch.from_numpy(clip_image_input["pixel_values"][0] ).unsqueeze(0 ).to(self.device ).half() __a = self.clip_model.get_image_features(__snake_case ) __a = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__snake_case ) __a = image_embeddings_clip.repeat_interleave(__snake_case , dim=0 ) return image_embeddings_clip @torch.enable_grad() def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = latents.detach().requires_grad_() __a = self.scheduler.scale_model_input(__snake_case , __snake_case ) # predict the noise residual __a = self.unet(__snake_case , __snake_case , encoder_hidden_states=__snake_case ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): __a = self.scheduler.alphas_cumprod[timestep] __a = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __a = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 __a = torch.sqrt(__snake_case ) __a = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , __snake_case ): __a = self.scheduler.sigmas[index] __a = latents - sigma * noise_pred else: raise ValueError(F"scheduler type {type(self.scheduler )} not supported" ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __a = 1 / 0.1_8215 * sample __a = self.vae.decode(__snake_case ).sample __a = (image / 2 + 0.5).clamp(0 , 1 ) __a = transforms.Resize(self.feature_extractor_size )(__snake_case ) __a = self.normalize(__snake_case ).to(latents.dtype ) __a = self.clip_model.get_image_features(__snake_case ) __a = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__snake_case ) __a = spherical_dist_loss(__snake_case , __snake_case ).mean() * clip_guidance_scale __a = -torch.autograd.grad(__snake_case , __snake_case )[0] if isinstance(self.scheduler , __snake_case ): __a = latents.detach() + grads * (sigma**2) __a = noise_pred_original else: __a = noise_pred_original - torch.sqrt(__snake_case ) * grads return noise_pred, latents @torch.no_grad() def __call__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = 512 , lowerCamelCase = 512 , lowerCamelCase = 0.6 , lowerCamelCase = 50 , lowerCamelCase = 7.5 , lowerCamelCase = 1 , lowerCamelCase = 0.0 , lowerCamelCase = 100 , lowerCamelCase = None , lowerCamelCase = "pil" , lowerCamelCase = True , lowerCamelCase = 0.8 , lowerCamelCase = 0.1 , lowerCamelCase = 0.1 , ): if isinstance(__snake_case , __snake_case ) and len(__snake_case ) != batch_size: raise ValueError(F"You have passed {batch_size} batch_size, but only {len(__snake_case )} generators." ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F"`height` and `width` have to be divisible by 8 but are {height} and {width}." ) if isinstance(__snake_case , torch.Generator ) and batch_size > 1: __a = [generator] + [None] * (batch_size - 1) __a = [ ('''model''', self.coca_model is None), ('''tokenizer''', self.coca_tokenizer is None), ('''transform''', self.coca_transform is None), ] __a = [x[0] for x in coca_is_none if x[1]] __a = ''', '''.join(__snake_case ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(__snake_case ): raise ValueError( F"Content prompt is None and CoCa [{coca_is_none_str}] is None." F"Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline." ) __a = self.get_image_description(__snake_case ) if style_prompt is None: if len(__snake_case ): raise ValueError( F"Style prompt is None and CoCa [{coca_is_none_str}] is None." F" Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline." ) __a = self.get_image_description(__snake_case ) # get prompt text embeddings for content and style __a = self.tokenizer( __snake_case , padding="max_length" , max_length=self.tokenizer.model_max_length , truncation=__snake_case , return_tensors="pt" , ) __a = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] __a = self.tokenizer( __snake_case , padding="max_length" , max_length=self.tokenizer.model_max_length , truncation=__snake_case , return_tensors="pt" , ) __a = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] __a = slerp(__snake_case , __snake_case , __snake_case ) # duplicate text embeddings for each generation per prompt __a = text_embeddings.repeat_interleave(__snake_case , dim=0 ) # set timesteps __a = '''offset''' in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) __a = {} if accepts_offset: __a = 1 self.scheduler.set_timesteps(__snake_case , **__snake_case ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) __a = self.get_timesteps(__snake_case , __snake_case , self.device ) __a = timesteps[:1].repeat(__snake_case ) # Preprocess image __a = preprocess(__snake_case , __snake_case , __snake_case ) __a = self.prepare_latents( __snake_case , __snake_case , __snake_case , text_embeddings.dtype , self.device , __snake_case ) __a = preprocess(__snake_case , __snake_case , __snake_case ) __a = self.prepare_latents( __snake_case , __snake_case , __snake_case , text_embeddings.dtype , self.device , __snake_case ) __a = slerp(__snake_case , __snake_case , __snake_case ) if clip_guidance_scale > 0: __a = self.get_clip_image_embeddings(__snake_case , __snake_case ) __a = self.get_clip_image_embeddings(__snake_case , __snake_case ) __a = slerp( __snake_case , __snake_case , __snake_case ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. __a = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: __a = content_text_input.input_ids.shape[-1] __a = self.tokenizer([""] , padding="max_length" , max_length=__snake_case , return_tensors="pt" ) __a = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt __a = uncond_embeddings.repeat_interleave(__snake_case , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __a = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. __a = (batch_size, self.unet.config.in_channels, height // 8, width // 8) __a = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps __a = torch.randn(__snake_case , generator=__snake_case , device="cpu" , dtype=__snake_case ).to( self.device ) else: __a = torch.randn(__snake_case , generator=__snake_case , device=self.device , dtype=__snake_case ) else: if latents.shape != latents_shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) __a = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler __a = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __a = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __a = {} if accepts_eta: __a = eta # check if the scheduler accepts generator __a = '''generator''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: __a = generator with self.progress_bar(total=__snake_case ): for i, t in enumerate(__snake_case ): # expand the latents if we are doing classifier free guidance __a = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __a = self.scheduler.scale_model_input(__snake_case , __snake_case ) # predict the noise residual __a = self.unet(__snake_case , __snake_case , encoder_hidden_states=__snake_case ).sample # perform classifier free guidance if do_classifier_free_guidance: __a = noise_pred.chunk(2 ) __a = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: __a = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) __a = self.cond_fn( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , ) # compute the previous noisy sample x_t -> x_t-1 __a = self.scheduler.step(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor __a = 1 / 0.1_8215 * latents __a = self.vae.decode(__snake_case ).sample __a = (image / 2 + 0.5).clamp(0 , 1 ) __a = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __a = self.numpy_to_pil(__snake_case ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=__snake_case , nsfw_content_detected=__snake_case )
705
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[Any] = { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json""", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case__ ( snake_case_ ): _snake_case : str = """blenderbot-small""" _snake_case : str = ["""past_key_values"""] _snake_case : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCamelCase=50265 , lowerCamelCase=512 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="gelu" , lowerCamelCase=512 , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , lowerCamelCase=1 , lowerCamelCase=False , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=2 , **lowerCamelCase , ): __a = vocab_size __a = max_position_embeddings __a = d_model __a = encoder_ffn_dim __a = encoder_layers __a = encoder_attention_heads __a = decoder_ffn_dim __a = decoder_layers __a = decoder_attention_heads __a = dropout __a = attention_dropout __a = activation_dropout __a = activation_function __a = init_std __a = encoder_layerdrop __a = decoder_layerdrop __a = use_cache __a = encoder_layers __a = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , is_encoder_decoder=lowerCamelCase , decoder_start_token_id=lowerCamelCase , forced_eos_token_id=lowerCamelCase , **lowerCamelCase , ) class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a = {0: "batch"} __a = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __a = {0: "batch", 1: "decoder_sequence"} __a = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} else: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = super().outputs else: __a = super(lowerCamelCase , self ).outputs if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Generate decoder inputs __a = seq_length if not self.use_past else 1 __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) __a = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} __a = dict(**lowerCamelCase , **lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape __a = common_inputs["decoder_input_ids"].shape[1] __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = decoder_seq_length + 3 __a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __a = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase )] , dim=1 ) __a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __a , __a = self.num_layers __a = min(lowerCamelCase , lowerCamelCase ) __a = max(lowerCamelCase , lowerCamelCase ) - min_num_layers __a = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), ) ) # TODO: test this. __a = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(lowerCamelCase , lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape # Not using the same length for past_key_values __a = seqlen + 2 __a , __a = self.num_layers __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = common_inputs["attention_mask"].dtype __a = torch.cat( [common_inputs["attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) __a = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(lowerCamelCase ) ] return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __a = tokenizer.num_special_tokens_to_add(lowerCamelCase ) __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence __a = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __a = dict(tokenizer(lowerCamelCase , return_tensors=lowerCamelCase ) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): if self.task in ["default", "seq2seq-lm"]: __a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) elif self.task == "causal-lm": __a = self._generate_dummy_inputs_for_causal_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) else: __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if self.task in ["default", "seq2seq-lm"]: __a = super()._flatten_past_key_values_(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __a = super(lowerCamelCase , self )._flatten_past_key_values_( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): _snake_case : int = ["input_features", "attention_mask"] def __init__( self , lowerCamelCase=80 , lowerCamelCase=16000 , lowerCamelCase=0.0 , lowerCamelCase=10 , lowerCamelCase=25 , lowerCamelCase="hamming_window" , lowerCamelCase=3_2768.0 , lowerCamelCase=0.97 , lowerCamelCase=1.0 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , **lowerCamelCase , ): super().__init__(feature_size=lowerCamelCase , sampling_rate=lowerCamelCase , padding_value=lowerCamelCase , **lowerCamelCase ) __a = feature_size __a = sampling_rate __a = padding_value __a = hop_length __a = win_length __a = frame_signal_scale __a = preemphasis_coeff __a = mel_floor __a = normalize_means __a = normalize_vars __a = win_function __a = return_attention_mask __a = win_length * sampling_rate // 1000 __a = hop_length * sampling_rate // 1000 __a = optimal_fft_length(self.sample_size ) __a = (self.n_fft // 2) + 1 def a__ ( self , lowerCamelCase ): if self.win_function == "hamming_window": __a = window_function(window_length=self.sample_size , name=self.win_function , periodic=lowerCamelCase ) else: __a = window_function(window_length=self.sample_size , name=self.win_function ) __a = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) __a = spectrogram( one_waveform * self.frame_signal_scale , window=lowerCamelCase , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=lowerCamelCase , preemphasis=self.preemphasis_coeff , mel_filters=lowerCamelCase , mel_floor=self.mel_floor , log_mel="log" , ) return msfc_features.T def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if self.normalize_means: __a = x[:input_length].mean(axis=0 ) __a = np.subtract(lowerCamelCase , lowerCamelCase ) if self.normalize_vars: __a = x[:input_length].std(axis=0 ) __a = np.divide(lowerCamelCase , lowerCamelCase ) if input_length < x.shape[0]: __a = padding_value # make sure array is in float32 __a = x.astype(np.floataa ) return x def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCamelCase , lowerCamelCase , self.padding_value ) for x, n in zip(lowerCamelCase , lowerCamelCase )] def __call__( self , lowerCamelCase , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" F" {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with" F" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) __a = isinstance(lowerCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) __a = is_batched_numpy or ( isinstance(lowerCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __a = [np.asarray(lowerCamelCase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCamelCase , np.ndarray ): __a = np.asarray(lowerCamelCase , dtype=np.floataa ) elif isinstance(lowerCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __a = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __a = [raw_speech] # extract fbank features __a = [self._extract_mfsc_features(lowerCamelCase ) for one_waveform in raw_speech] # convert into correct format for padding __a = BatchFeature({"input_features": features} ) __a = self.pad( lowerCamelCase , padding=lowerCamelCase , max_length=lowerCamelCase , truncation=lowerCamelCase , pad_to_multiple_of=lowerCamelCase , return_attention_mask=lowerCamelCase , **lowerCamelCase , ) # make sure list is in array format __a = padded_inputs.get("input_features" ) if isinstance(input_features[0] , lowerCamelCase ): __a = [np.asarray(lowerCamelCase , dtype=np.floataa ) for feature in input_features] __a = padded_inputs.get("attention_mask" ) if attention_mask is not None: __a = [np.asarray(lowerCamelCase , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: __a = ( np.array(lowerCamelCase , dtype=np.intaa ) if self._get_padding_strategies(lowerCamelCase , max_length=lowerCamelCase ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) __a = self.normalize( padded_inputs["input_features"] , attention_mask=lowerCamelCase ) if return_tensors is not None: __a = padded_inputs.convert_to_tensors(lowerCamelCase ) return padded_inputs
706
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=99 , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=9 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase=8 , lowerCamelCase=0.1 , lowerCamelCase=0.002 , lowerCamelCase=1 , lowerCamelCase=0 , lowerCamelCase=0 , lowerCamelCase=None , lowerCamelCase=None , ): __a = parent __a = batch_size __a = encoder_seq_length __a = decoder_seq_length # For common tests __a = self.decoder_seq_length __a = is_training __a = use_attention_mask __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = d_ff __a = relative_attention_num_buckets __a = dropout_rate __a = initializer_factor __a = eos_token_id __a = pad_token_id __a = decoder_start_token_id __a = None __a = decoder_layers def a__ ( self ): return TaConfig.from_pretrained("google/umt5-base" ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): if attention_mask is None: __a = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __a = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __a = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=lowerCamelCase ) if decoder_head_mask is None: __a = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) if cross_attn_head_mask is None: __a = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def a__ ( self ): __a = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __a = input_ids.clamp(self.pad_token_id + 1 ) __a = decoder_input_ids.clamp(self.pad_token_id + 1 ) __a = self.get_config() __a = config.num_attention_heads __a = self.prepare_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, input_dict def a__ ( self ): __a , __a = self.prepare_config_and_inputs() return config, inputs_dict def a__ ( self ): return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self ): return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase , attention_mask=lowerCamelCase , decoder_attention_mask=lowerCamelCase , ) __a = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ) __a = result.last_hidden_state __a = result.past_key_values __a = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).get_decoder().to(lowerCamelCase ).eval() # first forward pass __a = model(lowerCamelCase , use_cache=lowerCamelCase ) __a = model(lowerCamelCase ) __a = model(lowerCamelCase , use_cache=lowerCamelCase ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) + 1 ) __a , __a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = model(lowerCamelCase )["last_hidden_state"] __a = model(lowerCamelCase , past_key_values=lowerCamelCase )["last_hidden_state"] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -1, random_slice_idx].detach() __a = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).to(lowerCamelCase ).half().eval() __a = model(**lowerCamelCase )["last_hidden_state"] self.parent.assertFalse(torch.isnan(lowerCamelCase ).any().item() ) @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _snake_case : int = (UMTaForConditionalGeneration,) if is_torch_available() else () _snake_case : Optional[int] = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _snake_case : List[Any] = True _snake_case : Union[str, Any] = False _snake_case : Union[str, Any] = False _snake_case : Tuple = True _snake_case : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests _snake_case : Optional[Any] = [0.8, 0.9] def a__ ( self ): __a = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() __a = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=lowerCamelCase , opset_version=9 , input_names=["input_ids", "decoder_input_ids"] , ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase ) def a__ ( self ): __a = ["encoder_attentions", "decoder_attentions", "cross_attentions"] __a = self.model_tester.prepare_config_and_inputs() __a = config_and_inputs[0] __a = UMTaForConditionalGeneration(lowerCamelCase ).eval() model.to(lowerCamelCase ) __a = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=lowerCamelCase ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), } for attn_name, (name, mask) in zip(lowerCamelCase , head_masking.items() ): __a = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": __a = torch.ones( config.num_decoder_layers , config.num_heads , device=lowerCamelCase ) __a = model.generate( config_and_inputs[1]["input_ids"] , num_beams=1 , max_length=3 , output_attentions=lowerCamelCase , return_dict_in_generate=lowerCamelCase , **lowerCamelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step __a = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def a__ ( self ): pass @require_torch @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def a__ ( self ): __a = UMTaForConditionalGeneration.from_pretrained("google/umt5-small" , return_dict=lowerCamelCase ).to(lowerCamelCase ) __a = AutoTokenizer.from_pretrained("google/umt5-small" , use_fast=lowerCamelCase , legacy=lowerCamelCase ) __a = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] __a = tokenizer(lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ).input_ids # fmt: off __a = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase , lowerCamelCase ) __a = model.generate(input_ids.to(lowerCamelCase ) ) __a = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] __a = tokenizer.batch_decode(lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig __lowerCamelCase:str = { """facebook/maskformer-swin-base-ade""": ( """https://huggingface.co/facebook/maskformer-swin-base-ade/blob/main/config.json""" ) # See all MaskFormer models at https://huggingface.co/models?filter=maskformer } __lowerCamelCase:List[Any] = logging.get_logger(__name__) class snake_case__ ( _UpperCamelCase ): _snake_case : Tuple = "maskformer" _snake_case : Optional[int] = {"hidden_size": "mask_feature_size"} _snake_case : Optional[Any] = ["resnet", "swin"] _snake_case : Optional[int] = ["detr"] def __init__( self , lowerCamelCase = 256 , lowerCamelCase = 256 , lowerCamelCase = 0.1 , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = 0.02 , lowerCamelCase = 1.0 , lowerCamelCase = 1.0 , lowerCamelCase = 1.0 , lowerCamelCase = 20.0 , lowerCamelCase = None , **lowerCamelCase , ): if backbone_config is None: # fall back to https://huggingface.co/microsoft/swin-base-patch4-window12-384-in22k __a = SwinConfig( image_size=384 , in_channels=3 , patch_size=4 , embed_dim=128 , depths=[2, 2, 18, 2] , num_heads=[4, 8, 16, 32] , window_size=12 , drop_path_rate=0.3 , out_features=["stage1", "stage2", "stage3", "stage4"] , ) if isinstance(__a , __a ): __a = backbone_config.pop("model_type" ) __a = CONFIG_MAPPING[backbone_model_type] __a = config_class.from_dict(__a ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"Backbone {backbone_config.model_type} is not a supported model and may not be compatible with MaskFormer. " F"Supported model types: {','.join(self.backbones_supported )}" ) if decoder_config is None: # fall back to https://huggingface.co/facebook/detr-resnet-50 __a = DetrConfig() else: # verify that the decoder is supported __a = ( decoder_config.pop("model_type" ) if isinstance(__a , __a ) else decoder_config.model_type ) if decoder_type not in self.decoders_supported: raise ValueError( F"Transformer Decoder {decoder_type} not supported, please use one of" F" {','.join(self.decoders_supported )}" ) if isinstance(__a , __a ): __a = CONFIG_MAPPING[decoder_type] __a = config_class.from_dict(__a ) __a = backbone_config __a = decoder_config # main feature dimension for the model __a = fpn_feature_size __a = mask_feature_size # initializer __a = init_std __a = init_xavier_std # Hungarian matcher && loss __a = cross_entropy_weight __a = dice_weight __a = mask_weight __a = use_auxiliary_loss __a = no_object_weight __a = output_auxiliary_logits __a = self.decoder_config.encoder_attention_heads __a = self.decoder_config.num_hidden_layers super().__init__(**__a ) @classmethod def a__ ( cls , lowerCamelCase , lowerCamelCase , **lowerCamelCase ): return cls( backbone_config=__a , decoder_config=__a , **__a , ) def a__ ( self ): __a = copy.deepcopy(self.__dict__ ) __a = self.backbone_config.to_dict() __a = self.decoder_config.to_dict() __a = self.__class__.model_type return output
707
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase( a , a , a ): # Initialise PyTorch model __a = MobileBertConfig.from_json_file(a ) print(F"Building PyTorch model from configuration: {config}" ) __a = MobileBertForPreTraining(a ) # Load weights from tf checkpoint __a = load_tf_weights_in_mobilebert(a , a , a ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) SCREAMING_SNAKE_CASE__:List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
67
0
from __future__ import annotations class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase ): __a , __a = text, pattern __a , __a = len(_snake_case ), len(_snake_case ) def a__ ( self , lowerCamelCase ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def a__ ( self , lowerCamelCase ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def a__ ( self ): __a = [] for i in range(self.textLen - self.patLen + 1 ): __a = self.mismatch_in_text(_snake_case ) if mismatch_index == -1: positions.append(_snake_case ) else: __a = self.match_in_pattern(self.text[mismatch_index] ) __a = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions SCREAMING_SNAKE_CASE__:Optional[Any] = """ABAABA""" SCREAMING_SNAKE_CASE__:Optional[Any] = """AB""" SCREAMING_SNAKE_CASE__:int = BoyerMooreSearch(text, pattern) SCREAMING_SNAKE_CASE__:int = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
708
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__ ( snake_case_ ): def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __a = input_file.read() __a = regexp.search(lowerCamelCase ) return match def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __a = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __a = regexp.finditer(lowerCamelCase ) __a = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCamelCase ) ): raise AssertionError(F"open(...) must use utf-8 encoding in {dataset}" ) def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCamelCase ) ): raise AssertionError(F"print statement found in {dataset}. Use datasets.logger/logging instead." )
67
0
"""simple docstring""" from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def _lowerCamelCase( a ): def is_in_circle(a , a ) -> bool: __a = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle __a = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(a ) ) # The ratio of the area for circle to square is pi/4. __a = proportion * 4 print(F"The estimated value of pi is {pi_estimate}" ) print(F"The numpy value of pi is {pi}" ) print(F"The total error is {abs(pi - pi_estimate )}" ) def _lowerCamelCase( a , a , a = 0.0 , a = 1.0 , ): return mean( function_to_integrate(uniform(a , a ) ) for _ in range(a ) ) * (max_value - min_value) def _lowerCamelCase( a , a = 0.0 , a = 1.0 ): def identity_function(a ) -> float: return x __a = area_under_curve_estimator( a , a , a , a ) __a = (max_value * max_value - min_value * min_value) / 2 print("******************" ) print(F"Estimating area under y=x where x varies from {min_value} to {max_value}" ) print(F"Estimated value is {estimated_value}" ) print(F"Expected value is {expected_value}" ) print(F"Total error is {abs(estimated_value - expected_value )}" ) print("******************" ) def _lowerCamelCase( a ): def function_to_integrate(a ) -> float: return sqrt(4.0 - x * x ) __a = area_under_curve_estimator( a , a , 0.0 , 2.0 ) print("******************" ) print("Estimating pi using area_under_curve_estimator" ) print(F"Estimated value is {estimated_value}" ) print(F"Expected value is {pi}" ) print(F"Total error is {abs(estimated_value - pi )}" ) print("******************" ) if __name__ == "__main__": import doctest doctest.testmod()
709
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
67
0
"""simple docstring""" class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = None __a = None __a = graph self._normalize_graph(_a , _a ) __a = len(_a ) __a = None def a__ ( self , lowerCamelCase , lowerCamelCase ): if sources is int: __a = [sources] if sinks is int: __a = [sinks] if len(_a ) == 0 or len(_a ) == 0: return __a = sources[0] __a = sinks[0] # make fake vertex if there are more # than one source or sink if len(_a ) > 1 or len(_a ) > 1: __a = 0 for i in sources: max_input_flow += sum(self.graph[i] ) __a = len(self.graph ) + 1 for room in self.graph: room.insert(0 , 0 ) self.graph.insert(0 , [0] * size ) for i in sources: __a = max_input_flow __a = 0 __a = len(self.graph ) + 1 for room in self.graph: room.append(0 ) self.graph.append([0] * size ) for i in sinks: __a = max_input_flow __a = size - 1 def a__ ( self ): if self.maximum_flow_algorithm is None: raise Exception("You need to set maximum flow algorithm before." ) if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def a__ ( self , lowerCamelCase ): __a = algorithm(self ) class snake_case__ : def __init__( self , lowerCamelCase ): __a = flow_network __a = flow_network.verticesCount __a = flow_network.sourceIndex __a = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that __a = flow_network.graph __a = False def a__ ( self ): if not self.executed: self._algorithm() __a = True def a__ ( self ): pass class snake_case__ ( UpperCamelCase__ ): def __init__( self , lowerCamelCase ): super().__init__(_a ) # use this to save your result __a = -1 def a__ ( self ): if not self.executed: raise Exception("You should execute algorithm before using its result!" ) return self.maximum_flow class snake_case__ ( UpperCamelCase__ ): def __init__( self , lowerCamelCase ): super().__init__(_a ) __a = [[0] * self.verticies_count for i in range(self.verticies_count )] __a = [0] * self.verticies_count __a = [0] * self.verticies_count def a__ ( self ): __a = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index] ): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule __a = [ i for i in range(self.verticies_count ) if i != self.source_index and i != self.sink_index ] # move through list __a = 0 while i < len(_a ): __a = vertices_list[i] __a = self.heights[vertex_index] self.process_vertex(_a ) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(_a ) ) __a = 0 else: i += 1 __a = sum(self.preflow[self.source_index] ) def a__ ( self , lowerCamelCase ): while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count ): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(_a , _a ) self.relabel(_a ) def a__ ( self , lowerCamelCase , lowerCamelCase ): __a = min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def a__ ( self , lowerCamelCase ): __a = None for to_index in range(self.verticies_count ): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): __a = self.heights[to_index] if min_height is not None: __a = min_height + 1 if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Union[str, Any] = [0] SCREAMING_SNAKE_CASE__:Any = [3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] SCREAMING_SNAKE_CASE__:Tuple = [[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network SCREAMING_SNAKE_CASE__:Optional[Any] = FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate SCREAMING_SNAKE_CASE__:int = flow_network.find_maximum_flow() print(F'''maximum flow is {maximum_flow}''')
710
"""simple docstring""" import heapq import sys import numpy as np SCREAMING_SNAKE_CASE__:Optional[int] = tuple[int, int] class snake_case__ : def __init__( self ): __a = [] __a = set() def a__ ( self ): if not self.empty(): return self.elements[0][0] else: return float("inf" ) def a__ ( self ): return len(self.elements ) == 0 def a__ ( self , lowerCamelCase , lowerCamelCase ): if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(lowerCamelCase ) else: # update # print("update", item) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def a__ ( self , lowerCamelCase ): if item in self.set: self.set.remove(lowerCamelCase ) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def a__ ( self ): return self.elements[0][1] def a__ ( self ): ((__a) , (__a)) = heapq.heappop(self.elements ) self.set.remove(lowerCamelCase ) return (priority, item) def _lowerCamelCase( a , a ): # euclidean distance __a = np.array(a ) __a = np.array(a ) return np.linalg.norm(a - b ) def _lowerCamelCase( a , a ): # integer division by time variable return consistent_heuristic(a , a ) // t def _lowerCamelCase( a , a ): # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _lowerCamelCase( a , a , a , a ): __a = g_function[start] + Wa * heuristics[i](a , a ) return ans def _lowerCamelCase( a , a , a ): __a = np.chararray((n, n) ) for i in range(a ): for j in range(a ): __a = "*" for i in range(a ): for j in range(a ): if (j, (n - 1) - i) in blocks: __a = "#" __a = "-" __a = back_pointer[goal] while x != start: ((__a) , (__a)) = x # print(x) __a = "-" __a = back_pointer[x] __a = "-" for i in range(a ): for j in range(a ): if (i, j) == (0, n - 1): print(grid[i][j] , end=" " ) print("<-- End position" , end=" " ) else: print(grid[i][j] , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) print("PATH TAKEN BY THE ALGORITHM IS:-" ) __a = back_pointer[goal] while x != start: print(a , end=" " ) __a = back_pointer[x] print(a ) sys.exit() def _lowerCamelCase( a ): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _lowerCamelCase( a , a , a , a , a , a , a , a , ): for itera in range(a ): open_list[itera].remove_element(a ) # print("s", s) # print("j", j) ((__a) , (__a)) = s __a = (x - 1, y) __a = (x + 1, y) __a = (x, y + 1) __a = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(a ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(a ) __a = -1 __a = float("inf" ) if valid(a ) and g_function[neighbours] > g_function[s] + 1: __a = g_function[s] + 1 __a = s if neighbours not in close_list_anchor: open_list[0].put(a , key(a , 0 , a , a ) ) if neighbours not in close_list_inad: for var in range(1 , a ): if key(a , a , a , a ) <= Wa * key( a , 0 , a , a ): open_list[j].put( a , key(a , a , a , a ) ) def _lowerCamelCase( ): __a = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(1_5 , 2_0 ): some_list.append((x, 1_7) ) for x in range(1_0 , 1_9 ): for y in range(1 , 1_5 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(1_2 , 1_9 ): some_list.append((x, y) ) for x in range(3 , 1_3 ): for y in range(1_6 , 1_9 ): some_list.append((x, y) ) return some_list SCREAMING_SNAKE_CASE__:Any = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} SCREAMING_SNAKE_CASE__:str = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] SCREAMING_SNAKE_CASE__:int = make_common_ground() SCREAMING_SNAKE_CASE__:List[str] = blocks_blk # hyper parameters SCREAMING_SNAKE_CASE__:str = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 20 SCREAMING_SNAKE_CASE__:Dict = 3 # one consistent and two other inconsistent # start and end destination SCREAMING_SNAKE_CASE__:Dict = (0, 0) SCREAMING_SNAKE_CASE__:Optional[Any] = (n - 1, n - 1) SCREAMING_SNAKE_CASE__:List[str] = 1 def _lowerCamelCase( a , a , a ): __a = {start: 0, goal: float("inf" )} __a = {start: -1, goal: -1} __a = [] __a = set() for i in range(a ): open_list.append(PriorityQueue() ) open_list[i].put(a , key(a , a , a , a ) ) __a = [] __a = [] while open_list[0].minkey() < float("inf" ): for i in range(1 , a ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a , __a = open_list[i].top_show() visited.add(a ) expand_state( a , a , a , a , a , a , a , a , ) close_list_inad.append(a ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a = open_list[0].top_show() visited.add(a ) expand_state( a , 0 , a , a , a , a , a , a , ) close_list_anchor.append(a ) print("No path found to goal" ) print() for i in range(n - 1 , -1 , -1 ): for j in range(a ): if (j, i) in blocks: print("#" , end=" " ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" " ) else: print("-" , end=" " ) else: print("*" , end=" " ) if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
67
0
"""simple docstring""" import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): debug_launcher(test_script.main ) def a__ ( self ): debug_launcher(test_ops.main )
711
"""simple docstring""" SCREAMING_SNAKE_CASE__:Any = """Alexander Joslin""" import operator as op from .stack import Stack def _lowerCamelCase( a ): __a = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} __a = Stack() __a = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a ) ) elif i in operators: # RULE 2 operator_stack.push(a ) elif i == ")": # RULE 4 __a = operator_stack.peek() operator_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operators[opr](a , a ) operand_stack.push(a ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
67
0
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class snake_case__ ( unittest.TestCase ): def a__ ( self ): __a = "| <pad> <unk> <s> </s> a b c d e f g h i j k".split() __a = dict(zip(_lowerCamelCase , range(len(_lowerCamelCase ) ) ) ) __a = { "unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", } __a = { "feature_size": 1, "padding_value": 0.0, "sampling_rate": 16000, "return_attention_mask": False, "do_normalize": True, } __a = tempfile.mkdtemp() __a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __a = os.path.join(self.tmpdirname , _lowerCamelCase ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_lowerCamelCase ) + "\n" ) with open(self.feature_extraction_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_lowerCamelCase ) + "\n" ) # load decoder from hub __a = "hf-internal-testing/ngram-beam-search-decoder" def a__ ( self , **lowerCamelCase ): __a = self.add_kwargs_tokens_map.copy() kwargs.update(_lowerCamelCase ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def a__ ( self , **lowerCamelCase ): return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def a__ ( self , **lowerCamelCase ): return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **_lowerCamelCase ) def a__ ( self ): shutil.rmtree(self.tmpdirname ) def a__ ( self ): __a = self.get_tokenizer() __a = self.get_feature_extractor() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) __a = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , _lowerCamelCase ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , _lowerCamelCase ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , _lowerCamelCase ) def a__ ( self ): __a = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match __a = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def a__ ( self ): __a = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["xx"] ) with self.assertRaisesRegex(_lowerCamelCase , "include" ): WavaVecaProcessorWithLM( tokenizer=_lowerCamelCase , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def a__ ( self ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) __a = floats_list((3, 1000) ) __a = feature_extractor(_lowerCamelCase , return_tensors="np" ) __a = processor(_lowerCamelCase , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def a__ ( self ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) __a = "This is a test string" __a = processor(text=_lowerCamelCase ) __a = tokenizer(_lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def a__ ( self , lowerCamelCase=(2, 10, 16) , lowerCamelCase=77 ): np.random.seed(_lowerCamelCase ) return np.random.rand(*_lowerCamelCase ) def a__ ( self ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) __a = self._get_dummy_logits(shape=(10, 16) , seed=13 ) __a = processor.decode(_lowerCamelCase ) __a = decoder.decode_beams(_lowerCamelCase )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual("</s> <s> </s>" , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ["fork"], ["spawn"]] ) def a__ ( self , lowerCamelCase ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) __a = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: __a = processor.batch_decode(_lowerCamelCase ) else: with get_context(_lowerCamelCase ).Pool() as pool: __a = processor.batch_decode(_lowerCamelCase , _lowerCamelCase ) __a = list(_lowerCamelCase ) with get_context("fork" ).Pool() as p: __a = decoder.decode_beams_batch(_lowerCamelCase , _lowerCamelCase ) __a , __a , __a = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(_lowerCamelCase , decoded_processor.text ) self.assertListEqual(["<s> <s> </s>", "<s> <s> <s>"] , decoded_processor.text ) self.assertListEqual(_lowerCamelCase , decoded_processor.logit_score ) self.assertListEqual(_lowerCamelCase , decoded_processor.lm_score ) def a__ ( self ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) __a = self._get_dummy_logits() __a = 15 __a = -20.0 __a = -4.0 __a = processor.batch_decode( _lowerCamelCase , beam_width=_lowerCamelCase , beam_prune_logp=_lowerCamelCase , token_min_logp=_lowerCamelCase , ) __a = decoded_processor_out.text __a = list(_lowerCamelCase ) with get_context("fork" ).Pool() as pool: __a = decoder.decode_beams_batch( _lowerCamelCase , _lowerCamelCase , beam_width=_lowerCamelCase , beam_prune_logp=_lowerCamelCase , token_min_logp=_lowerCamelCase , ) __a = [d[0][0] for d in decoded_decoder_out] __a = [d[0][2] for d in decoded_decoder_out] __a = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) self.assertListEqual(["</s> <s> <s>", "<s> <s> <s>"] , _lowerCamelCase ) self.assertTrue(np.array_equal(_lowerCamelCase , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , _lowerCamelCase , atol=1E-3 ) ) self.assertTrue(np.array_equal(_lowerCamelCase , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , _lowerCamelCase , atol=1E-3 ) ) def a__ ( self ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) __a = self._get_dummy_logits() __a = 2.0 __a = 5.0 __a = -20.0 __a = True __a = processor.batch_decode( _lowerCamelCase , alpha=_lowerCamelCase , beta=_lowerCamelCase , unk_score_offset=_lowerCamelCase , lm_score_boundary=_lowerCamelCase , ) __a = decoded_processor_out.text __a = list(_lowerCamelCase ) decoder.reset_params( alpha=_lowerCamelCase , beta=_lowerCamelCase , unk_score_offset=_lowerCamelCase , lm_score_boundary=_lowerCamelCase , ) with get_context("fork" ).Pool() as pool: __a = decoder.decode_beams_batch( _lowerCamelCase , _lowerCamelCase , ) __a = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) self.assertListEqual(["<s> </s> <s> </s> </s>", "</s> </s> <s> </s> </s>"] , _lowerCamelCase ) __a = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , _lowerCamelCase ) def a__ ( self ): __a = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) __a = processor.decoder.model_container[processor.decoder._model_key] __a = Path(language_model._kenlm_model.path.decode("utf-8" ) ).parent.parent.absolute() __a = os.listdir(_lowerCamelCase ) __a = ["alphabet.json", "language_model"] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def a__ ( self ): __a = snapshot_download("hf-internal-testing/processor_with_lm" ) __a = WavaVecaProcessorWithLM.from_pretrained(_lowerCamelCase ) __a = processor.decoder.model_container[processor.decoder._model_key] __a = Path(language_model._kenlm_model.path.decode("utf-8" ) ).parent.parent.absolute() __a = os.listdir(_lowerCamelCase ) __a = os.listdir(_lowerCamelCase ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def a__ ( self ): __a = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) __a = AutoProcessor.from_pretrained("hf-internal-testing/processor_with_lm" ) __a = floats_list((3, 1000) ) __a = processor_wavaveca(_lowerCamelCase , return_tensors="np" ) __a = processor_auto(_lowerCamelCase , return_tensors="np" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) __a = self._get_dummy_logits() __a = processor_wavaveca.batch_decode(_lowerCamelCase ) __a = processor_auto.batch_decode(_lowerCamelCase ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def a__ ( self ): __a = self.get_feature_extractor() __a = self.get_tokenizer() __a = self.get_decoder() __a = WavaVecaProcessorWithLM(tokenizer=_lowerCamelCase , feature_extractor=_lowerCamelCase , decoder=_lowerCamelCase ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , ) @staticmethod def a__ ( lowerCamelCase , lowerCamelCase ): __a = [d[key] for d in offsets] return retrieved_list def a__ ( self ): __a = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) __a = self._get_dummy_logits()[0] __a = processor.decode(_lowerCamelCase , output_word_offsets=_lowerCamelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("text" in outputs ) self.assertTrue("word_offsets" in outputs ) self.assertTrue(isinstance(_lowerCamelCase , _lowerCamelCase ) ) self.assertEqual(" ".join(self.get_from_offsets(outputs["word_offsets"] , "word" ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "word" ) , ["<s>", "<s>", "</s>"] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "start_offset" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"] , "end_offset" ) , [1, 3, 5] ) def a__ ( self ): __a = WavaVecaProcessorWithLM.from_pretrained("hf-internal-testing/processor_with_lm" ) __a = self._get_dummy_logits() __a = processor.batch_decode(_lowerCamelCase , output_word_offsets=_lowerCamelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("text" in outputs ) self.assertTrue("word_offsets" in outputs ) self.assertTrue(isinstance(_lowerCamelCase , _lowerCamelCase ) ) self.assertListEqual( [" ".join(self.get_from_offsets(_lowerCamelCase , "word" ) ) for o in outputs["word_offsets"]] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "word" ) , ["<s>", "<s>", "</s>"] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "start_offset" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["word_offsets"][0] , "end_offset" ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def a__ ( self ): import torch __a = load_dataset("common_voice" , "en" , split="train" , streaming=_lowerCamelCase ) __a = ds.cast_column("audio" , datasets.Audio(sampling_rate=16000 ) ) __a = iter(_lowerCamelCase ) __a = next(_lowerCamelCase ) __a = AutoProcessor.from_pretrained("patrickvonplaten/wav2vec2-base-100h-with-lm" ) __a = WavaVecaForCTC.from_pretrained("patrickvonplaten/wav2vec2-base-100h-with-lm" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train __a = processor(sample["audio"]["array"] , return_tensors="pt" ).input_values with torch.no_grad(): __a = model(_lowerCamelCase ).logits.cpu().numpy() __a = processor.decode(logits[0] , output_word_offsets=_lowerCamelCase ) __a = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate __a = [ { "start_time": d["start_offset"] * time_offset, "end_time": d["end_offset"] * time_offset, "word": d["word"], } for d in output["word_offsets"] ] __a = "WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL" # output words self.assertEqual(" ".join(self.get_from_offsets(_lowerCamelCase , "word" ) ) , _lowerCamelCase ) self.assertEqual(" ".join(self.get_from_offsets(_lowerCamelCase , "word" ) ) , output.text ) # output times __a = torch.tensor(self.get_from_offsets(_lowerCamelCase , "start_time" ) ) __a = torch.tensor(self.get_from_offsets(_lowerCamelCase , "end_time" ) ) # fmt: off __a = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) __a = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=0.01 ) ) self.assertTrue(torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=0.01 ) )
712
"""simple docstring""" from math import pi def _lowerCamelCase( a , a ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
67
0
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class snake_case__ ( pl.LightningModule ): def __init__( self , lowerCamelCase ): super().__init__() __a = model __a = 2 __a = nn.Linear(self.model.config.hidden_size , self.num_labels ) def a__ ( self ): pass def _lowerCamelCase( a , a , a ): __a = LongformerModel.from_pretrained(__snake_case ) __a = LightningModel(__snake_case ) __a = torch.load(__snake_case , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model __a = LongformerForQuestionAnswering.from_pretrained(__snake_case ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(__snake_case ) print(F"Conversion successful. Model saved under {pytorch_dump_folder_path}" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--longformer_model""", default=None, type=str, required=True, help="""model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.""", ) parser.add_argument( """--longformer_question_answering_ckpt_path""", default=None, type=str, required=True, help="""Path the official PyTorch Lightning Checkpoint.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
713
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): _snake_case : Dict = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = IMAGENET_DEFAULT_MEAN , lowerCamelCase = IMAGENET_DEFAULT_STD , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __a = image_std if image_std is not None else IMAGENET_DEFAULT_STD def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __a = int((256 / 224) * size["shortest_edge"] ) __a = get_resize_output_image_size(lowerCamelCase , size=lowerCamelCase , default_to_square=lowerCamelCase ) __a = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}" ) return resize( lowerCamelCase , size=(size_dict["height"], size_dict["width"]) , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"Size dict must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(lowerCamelCase , lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(lowerCamelCase , lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
"""simple docstring""" import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Any = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) # !wget https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml parser.add_argument( """--original_config_file""", default=None, type=str, help="""The YAML config file corresponding to the original architecture.""", ) parser.add_argument( """--num_in_channels""", default=None, type=int, help="""The number of input channels. If `None` number of input channels will be automatically inferred.""", ) parser.add_argument( """--scheduler_type""", default="""pndm""", type=str, help="""Type of scheduler to use. Should be one of [\'pndm\', \'lms\', \'ddim\', \'euler\', \'euler-ancestral\', \'dpm\']""", ) parser.add_argument( """--pipeline_type""", default=None, type=str, help=( """The pipeline type. One of \'FrozenOpenCLIPEmbedder\', \'FrozenCLIPEmbedder\', \'PaintByExample\'""" """. If `None` pipeline will be automatically inferred.""" ), ) parser.add_argument( """--image_size""", default=None, type=int, help=( """The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2""" """ Base. Use 768 for Stable Diffusion v2.""" ), ) parser.add_argument( """--prediction_type""", default=None, type=str, help=( """The prediction type that the model was trained on. Use \'epsilon\' for Stable Diffusion v1.X and Stable""" """ Diffusion v2 Base. Use \'v_prediction\' for Stable Diffusion v2.""" ), ) parser.add_argument( """--extract_ema""", action="""store_true""", help=( """Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights""" """ or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield""" """ higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.""" ), ) parser.add_argument( """--upcast_attention""", action="""store_true""", help=( """Whether the attention computation should always be upcasted. This is necessary when running stable""" """ diffusion 2.1.""" ), ) parser.add_argument( """--from_safetensors""", action="""store_true""", help="""If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.""", ) parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") parser.add_argument( """--stable_unclip""", type=str, default=None, required=False, help="""Set if this is a stable unCLIP model. One of \'txt2img\' or \'img2img\'.""", ) parser.add_argument( """--stable_unclip_prior""", type=str, default=None, required=False, help="""Set if this is a stable unCLIP txt2img model. Selects which prior to use. If `--stable_unclip` is set to `txt2img`, the karlo prior (https://huggingface.co/kakaobrain/karlo-v1-alpha/tree/main/prior) is selected by default.""", ) parser.add_argument( """--clip_stats_path""", type=str, help="""Path to the clip stats file. Only required if the stable unclip model\'s config specifies `model.params.noise_aug_config.params.clip_stats_path`.""", required=False, ) parser.add_argument( """--controlnet""", action="""store_true""", default=None, help="""Set flag if this is a controlnet checkpoint.""" ) parser.add_argument("""--half""", action="""store_true""", help="""Save weights in half precision.""") parser.add_argument( """--vae_path""", type=str, default=None, required=False, help="""Set to a path, hub id to an already converted vae to not convert it again.""", ) SCREAMING_SNAKE_CASE__:List[str] = parser.parse_args() SCREAMING_SNAKE_CASE__:Optional[Any] = download_from_original_stable_diffusion_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, prediction_type=args.prediction_type, model_type=args.pipeline_type, extract_ema=args.extract_ema, scheduler_type=args.scheduler_type, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, stable_unclip=args.stable_unclip, stable_unclip_prior=args.stable_unclip_prior, clip_stats_path=args.clip_stats_path, controlnet=args.controlnet, vae_path=args.vae_path, ) if args.half: pipe.to(torch_dtype=torch.floataa) if args.controlnet: # only save the controlnet model pipe.controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors) else: pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
714
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=30 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=None , lowerCamelCase=2 , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def a__ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, pixel_values, labels def a__ ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTForMaskedImageModeling(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __a = 1 __a = ViTForMaskedImageModeling(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.type_sequence_label_size __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a = 1 __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, unittest.TestCase ): _snake_case : Any = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) _snake_case : List[Any] = ( {"""feature-extraction""": ViTModel, """image-classification""": ViTForImageClassification} if is_torch_available() else {} ) _snake_case : int = True _snake_case : int = False _snake_case : str = False _snake_case : Optional[Any] = False def a__ ( self ): __a = ViTModelTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def a__ ( self ): pass def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def a__ ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def _lowerCamelCase( ): __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def a__ ( self ): return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def a__ ( self ): __a = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(lowerCamelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(**lowerCamelCase ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __a = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def a__ ( self ): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. __a = ViTModel.from_pretrained("facebook/dino-vits8" ).to(lowerCamelCase ) __a = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(lowerCamelCase , interpolate_pos_encoding=lowerCamelCase ) # verify the logits __a = torch.Size((1, 3601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , lowerCamelCase ) __a = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def a__ ( self ): __a = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __a = model(lowerCamelCase )
67
0
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class snake_case__ ( UpperCamelCase__, unittest.TestCase ): _snake_case : int = ShapEPipeline _snake_case : Any = ["""prompt"""] _snake_case : int = ["""prompt"""] _snake_case : Any = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] _snake_case : List[str] = False @property def a__ ( self ): return 32 @property def a__ ( self ): return 32 @property def a__ ( self ): return self.time_input_dim * 4 @property def a__ ( self ): return 8 @property def a__ ( self ): __a = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) return tokenizer @property def a__ ( self ): torch.manual_seed(0 ) __a = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(__A ) @property def a__ ( self ): torch.manual_seed(0 ) __a = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } __a = PriorTransformer(**__A ) return model @property def a__ ( self ): torch.manual_seed(0 ) __a = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } __a = ShapERenderer(**__A ) return model def a__ ( self ): __a = self.dummy_prior __a = self.dummy_text_encoder __a = self.dummy_tokenizer __a = self.dummy_renderer __a = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1024 , prediction_type="sample" , use_karras_sigmas=__A , clip_sample=__A , clip_sample_range=1.0 , ) __a = { "prior": prior, "text_encoder": text_encoder, "tokenizer": tokenizer, "renderer": renderer, "scheduler": scheduler, } return components def a__ ( self , lowerCamelCase , lowerCamelCase=0 ): if str(__A ).startswith("mps" ): __a = torch.manual_seed(__A ) else: __a = torch.Generator(device=__A ).manual_seed(__A ) __a = { "prompt": "horse", "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def a__ ( self ): __a = "cpu" __a = self.get_dummy_components() __a = self.pipeline_class(**__A ) __a = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) __a = pipe(**self.get_dummy_inputs(__A ) ) __a = output.images[0] __a = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __a = np.array( [ 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def a__ ( self ): self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def a__ ( self ): __a = torch_device == "cpu" __a = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=__A , relax_max_difference=__A , ) def a__ ( self ): __a = self.get_dummy_components() __a = self.pipeline_class(**__A ) __a = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) __a = 1 __a = 2 __a = self.get_dummy_inputs(__A ) for key in inputs.keys(): if key in self.batch_params: __a = batch_size * [inputs[key]] __a = pipe(**__A , num_images_per_prompt=__A )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_np_out.npy" ) __a = ShapEPipeline.from_pretrained("openai/shap-e" ) __a = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) __a = torch.Generator(device=__A ).manual_seed(0 ) __a = pipe( "a shark" , generator=__A , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(__A , __A )
715
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=True , lowerCamelCase=1 / 255 , lowerCamelCase=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __a = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __a = parent __a = batch_size __a = num_channels __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std __a = do_rescale __a = rescale_factor __a = do_pad def a__ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def a__ ( self , lowerCamelCase , lowerCamelCase=False ): if not batched: __a = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __a , __a = image.size else: __a , __a = image.shape[1], image.shape[2] if w < h: __a = int(self.size["shortest_edge"] * h / w ) __a = self.size["shortest_edge"] elif w > h: __a = self.size["shortest_edge"] __a = int(self.size["shortest_edge"] * w / h ) else: __a = self.size["shortest_edge"] __a = self.size["shortest_edge"] else: __a = [] for image in image_inputs: __a , __a = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : List[Any] = DetaImageProcessor if is_vision_available() else None def a__ ( self ): __a = DetaImageProcessingTester(self ) @property def a__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_rescale" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_pad" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def a__ ( self ): __a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def a__ ( self ): pass def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self ): # prepare image and target __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"image_id": 39769, "annotations": target} # encode them __a = DetaImageProcessor() __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def a__ ( self ): # prepare image, target and masks_path __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __a = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __a = DetaImageProcessor(format="coco_panoptic" ) __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __a = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
67
0
"""simple docstring""" # A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def _lowerCamelCase( a ): __a = [False] * len(_UpperCAmelCase ) __a = [-1] * len(_UpperCAmelCase ) def dfs(a , a ): __a = True __a = c for u in graph[v]: if not visited[u]: dfs(_UpperCAmelCase , 1 - c ) for i in range(len(_UpperCAmelCase ) ): if not visited[i]: dfs(_UpperCAmelCase , 0 ) for i in range(len(_UpperCAmelCase ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph SCREAMING_SNAKE_CASE__:int = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
716
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE__:Dict = logging.getLogger() def _lowerCamelCase( ): __a = argparse.ArgumentParser() parser.add_argument("-f" ) __a = parser.parse_args() return args.f class snake_case__ ( snake_case_ ): def a__ ( self ): __a = logging.StreamHandler(sys.stdout ) logger.addHandler(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowerCamelCase , "argv" , lowerCamelCase ): __a = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self ): __a = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase )
67
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE__:List[Any] = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Optional[int] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Any = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys SCREAMING_SNAKE_CASE__:List[str] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
717
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __a = image_std if image_std is not None else OPENAI_CLIP_STD __a = do_convert_rgb def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) __a = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , param_name="size" , default_to_square=lowerCamelCase ) __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" , default_to_square=lowerCamelCase ) __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __a = [convert_to_rgb(lowerCamelCase ) for image in images] # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def _lowerCamelCase( ): __a = "https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png" __a = Image.open(requests.get(_lowercase , stream=_lowercase ).raw ).convert("RGB" ) return image def _lowerCamelCase( a ): __a = [] # fmt: off # vision encoder rename_keys.append(("visual_encoder.cls_token", "vision_model.embeddings.class_embedding") ) rename_keys.append(("visual_encoder.pos_embed", "vision_model.embeddings.position_embedding") ) rename_keys.append(("visual_encoder.patch_embed.proj.weight", "vision_model.embeddings.patch_embedding.weight") ) rename_keys.append(("visual_encoder.patch_embed.proj.bias", "vision_model.embeddings.patch_embedding.bias") ) rename_keys.append(("ln_vision.weight", "vision_model.post_layernorm.weight") ) rename_keys.append(("ln_vision.bias", "vision_model.post_layernorm.bias") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F"visual_encoder.blocks.{i}.norm1.weight", F"vision_model.encoder.layers.{i}.layer_norm1.weight") ) rename_keys.append((F"visual_encoder.blocks.{i}.norm1.bias", F"vision_model.encoder.layers.{i}.layer_norm1.bias") ) rename_keys.append((F"visual_encoder.blocks.{i}.norm2.weight", F"vision_model.encoder.layers.{i}.layer_norm2.weight") ) rename_keys.append((F"visual_encoder.blocks.{i}.norm2.bias", F"vision_model.encoder.layers.{i}.layer_norm2.bias") ) rename_keys.append((F"visual_encoder.blocks.{i}.attn.qkv.weight", F"vision_model.encoder.layers.{i}.self_attn.qkv.weight") ) rename_keys.append((F"visual_encoder.blocks.{i}.attn.proj.weight", F"vision_model.encoder.layers.{i}.self_attn.projection.weight",) ) rename_keys.append((F"visual_encoder.blocks.{i}.attn.proj.bias", F"vision_model.encoder.layers.{i}.self_attn.projection.bias") ) rename_keys.append((F"visual_encoder.blocks.{i}.mlp.fc1.weight", F"vision_model.encoder.layers.{i}.mlp.fc1.weight") ) rename_keys.append((F"visual_encoder.blocks.{i}.mlp.fc1.bias", F"vision_model.encoder.layers.{i}.mlp.fc1.bias") ) rename_keys.append((F"visual_encoder.blocks.{i}.mlp.fc2.weight", F"vision_model.encoder.layers.{i}.mlp.fc2.weight") ) rename_keys.append((F"visual_encoder.blocks.{i}.mlp.fc2.bias", F"vision_model.encoder.layers.{i}.mlp.fc2.bias") ) # QFormer rename_keys.append(("Qformer.bert.embeddings.LayerNorm.weight", "qformer.layernorm.weight") ) rename_keys.append(("Qformer.bert.embeddings.LayerNorm.bias", "qformer.layernorm.bias") ) # fmt: on return rename_keys def _lowerCamelCase( a , a , a ): __a = dct.pop(_lowercase ) __a = val def _lowerCamelCase( a , a ): for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases __a = state_dict.pop(F"visual_encoder.blocks.{i}.attn.q_bias" ) __a = state_dict.pop(F"visual_encoder.blocks.{i}.attn.v_bias" ) # next, set bias in the state dict __a = torch.cat((q_bias, torch.zeros_like(_lowercase , requires_grad=_lowercase ), v_bias) ) __a = qkv_bias def _lowerCamelCase( a , a ): __a = 3_6_4 if "coco" in model_name else 2_2_4 __a = BlipaVisionConfig(image_size=_lowercase ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: __a = OPTConfig.from_pretrained("facebook/opt-2.7b" , eos_token_id=_lowercase ).to_dict() elif "opt-6.7b" in model_name: __a = OPTConfig.from_pretrained("facebook/opt-6.7b" , eos_token_id=_lowercase ).to_dict() elif "t5-xl" in model_name: __a = TaConfig.from_pretrained("google/flan-t5-xl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: __a = TaConfig.from_pretrained("google/flan-t5-xxl" , dense_act_fn="gelu" , bos_token_id=1 ).to_dict() __a = BlipaConfig(vision_config=_lowercase , text_config=_lowercase ) return config, image_size @torch.no_grad() def _lowerCamelCase( a , a=None , a=False ): __a = ( AutoTokenizer.from_pretrained("facebook/opt-2.7b" ) if "opt" in model_name else AutoTokenizer.from_pretrained("google/flan-t5-xl" ) ) __a = tokenizer("\n" , add_special_tokens=_lowercase ).input_ids[0] __a , __a = get_blipa_config(_lowercase , eos_token_id=_lowercase ) __a = BlipaForConditionalGeneration(_lowercase ).eval() __a = { "blip2-opt-2.7b": ("blip2_opt", "pretrain_opt2.7b"), "blip2-opt-6.7b": ("blip2_opt", "pretrain_opt6.7b"), "blip2-opt-2.7b-coco": ("blip2_opt", "caption_coco_opt2.7b"), "blip2-opt-6.7b-coco": ("blip2_opt", "caption_coco_opt6.7b"), "blip2-flan-t5-xl": ("blip2_t5", "pretrain_flant5xl"), "blip2-flan-t5-xl-coco": ("blip2_t5", "caption_coco_flant5xl"), "blip2-flan-t5-xxl": ("blip2_t5", "pretrain_flant5xxl"), } __a , __a = model_name_to_original[model_name] # load original model print("Loading original model..." ) __a = "cuda" if torch.cuda.is_available() else "cpu" __a , __a , __a = load_model_and_preprocess( name=_lowercase , model_type=_lowercase , is_eval=_lowercase , device=_lowercase ) original_model.eval() print("Done!" ) # update state dict keys __a = original_model.state_dict() __a = create_rename_keys(_lowercase ) for src, dest in rename_keys: rename_key(_lowercase , _lowercase , _lowercase ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): __a = state_dict.pop(_lowercase ) if key.startswith("Qformer.bert" ): __a = key.replace("Qformer.bert" , "qformer" ) if "attention.self" in key: __a = key.replace("self" , "attention" ) if "opt_proj" in key: __a = key.replace("opt_proj" , "language_projection" ) if "t5_proj" in key: __a = key.replace("t5_proj" , "language_projection" ) if key.startswith("opt" ): __a = key.replace("opt" , "language" ) if key.startswith("t5" ): __a = key.replace("t5" , "language" ) __a = val # read in qv biases read_in_q_v_bias(_lowercase , _lowercase ) __a , __a = hf_model.load_state_dict(_lowercase , strict=_lowercase ) assert len(_lowercase ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] __a = load_demo_image() __a = vis_processors["eval"](_lowercase ).unsqueeze(0 ).to(_lowercase ) __a = tokenizer(["\n"] , return_tensors="pt" ).input_ids.to(_lowercase ) # create processor __a = BlipImageProcessor( size={"height": image_size, "width": image_size} , image_mean=_lowercase , image_std=_lowercase ) __a = BlipaProcessor(image_processor=_lowercase , tokenizer=_lowercase ) __a = processor(images=_lowercase , return_tensors="pt" ).pixel_values.to(_lowercase ) # make sure processor creates exact same pixel values assert torch.allclose(_lowercase , _lowercase ) original_model.to(_lowercase ) hf_model.to(_lowercase ) with torch.no_grad(): if "opt" in model_name: __a = original_model({"image": original_pixel_values, "text_input": [""]} ).logits __a = hf_model(_lowercase , _lowercase ).logits else: __a = original_model( {"image": original_pixel_values, "text_input": ["\n"], "text_output": ["\n"]} ).logits __a = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -1_0_0 ) __a = hf_model(_lowercase , _lowercase , labels=_lowercase ).logits assert original_logits.shape == logits.shape print("First values of original logits:" , original_logits[0, :3, :3] ) print("First values of HF logits:" , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": __a = torch.tensor( [[-41.58_50, -4.44_40, -8.99_22], [-47.43_22, -5.91_43, -1.73_40]] , device=_lowercase ) assert torch.allclose(logits[0, :3, :3] , _lowercase , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": __a = torch.tensor( [[-57.01_09, -9.89_67, -12.62_80], [-68.65_78, -12.71_91, -10.50_65]] , device=_lowercase ) else: # cast to same type __a = logits.dtype assert torch.allclose(original_logits.to(_lowercase ) , _lowercase , atol=1E-2 ) print("Looks ok!" ) print("Generating a caption..." ) __a = "" __a = tokenizer(_lowercase , return_tensors="pt" ).input_ids.to(_lowercase ) __a = original_model.generate({"image": original_pixel_values} ) __a = hf_model.generate( _lowercase , _lowercase , do_sample=_lowercase , num_beams=5 , max_length=3_0 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print("Original generation:" , _lowercase ) __a = input_ids.shape[1] __a = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=_lowercase ) __a = [text.strip() for text in output_text] print("HF generation:" , _lowercase ) if pytorch_dump_folder_path is not None: processor.save_pretrained(_lowercase ) hf_model.save_pretrained(_lowercase ) if push_to_hub: processor.push_to_hub(F"nielsr/{model_name}" ) hf_model.push_to_hub(F"nielsr/{model_name}" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Union[str, Any] = argparse.ArgumentParser() SCREAMING_SNAKE_CASE__:Tuple = [ """blip2-opt-2.7b""", """blip2-opt-6.7b""", """blip2-opt-2.7b-coco""", """blip2-opt-6.7b-coco""", """blip2-flan-t5-xl""", """blip2-flan-t5-xl-coco""", """blip2-flan-t5-xxl""", ] parser.add_argument( """--model_name""", default="""blip2-opt-2.7b""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) SCREAMING_SNAKE_CASE__:Dict = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
718
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Any = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Optional[Any] = { """vocab_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/vocab.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/vocab.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/vocab.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/vocab.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/vocab.json""", }, """merges_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/merges.txt""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/merges.txt""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/merges.txt""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/merges.txt""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/merges.txt""", }, """tokenizer_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/tokenizer.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/tokenizer.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__:Union[str, Any] = { """gpt2""": 1024, """gpt2-medium""": 1024, """gpt2-large""": 1024, """gpt2-xl""": 1024, """distilgpt2""": 1024, } class snake_case__ ( snake_case_ ): _snake_case : Tuple = VOCAB_FILES_NAMES _snake_case : str = PRETRAINED_VOCAB_FILES_MAP _snake_case : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : List[str] = ["""input_ids""", """attention_mask"""] _snake_case : Dict = GPTaTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , add_prefix_space=lowerCamelCase , **lowerCamelCase , ) __a = kwargs.pop("add_bos_token" , lowerCamelCase ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def _lowerCamelCase( a , a ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def _lowerCamelCase( a , a , a ): __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __a = ParquetDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase , keep_in_memory=_lowerCamelCase ).read() _check_parquet_dataset(_lowerCamelCase , _lowerCamelCase ) @pytest.mark.parametrize( "features" , [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ] , ) def _lowerCamelCase( a , a , a ): __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __a = features.copy() if features else default_expected_features __a = ( Features({feature: Value(_lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __a = ParquetDatasetReader(_lowerCamelCase , features=_lowerCamelCase , cache_dir=_lowerCamelCase ).read() _check_parquet_dataset(_lowerCamelCase , _lowerCamelCase ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def _lowerCamelCase( a , a , a ): __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __a = ParquetDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase , split=_lowerCamelCase ).read() _check_parquet_dataset(_lowerCamelCase , _lowerCamelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def _lowerCamelCase( a , a , a ): if issubclass(_lowerCamelCase , _lowerCamelCase ): __a = parquet_path elif issubclass(_lowerCamelCase , _lowerCamelCase ): __a = [parquet_path] __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __a = ParquetDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase ).read() _check_parquet_dataset(_lowerCamelCase , _lowerCamelCase ) def _lowerCamelCase( a , a , a=("train",) ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) for split in splits: __a = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def _lowerCamelCase( a , a , a ): __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __a = ParquetDatasetReader( {"train": parquet_path} , cache_dir=_lowerCamelCase , keep_in_memory=_lowerCamelCase ).read() _check_parquet_datasetdict(_lowerCamelCase , _lowerCamelCase ) @pytest.mark.parametrize( "features" , [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ] , ) def _lowerCamelCase( a , a , a ): __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __a = features.copy() if features else default_expected_features __a = ( Features({feature: Value(_lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) __a = ParquetDatasetReader({"train": parquet_path} , features=_lowerCamelCase , cache_dir=_lowerCamelCase ).read() _check_parquet_datasetdict(_lowerCamelCase , _lowerCamelCase ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def _lowerCamelCase( a , a , a ): if split: __a = {split: parquet_path} else: __a = """train""" __a = {"""train""": parquet_path, """test""": parquet_path} __a = tmp_path / """cache""" __a = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} __a = ParquetDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase ).read() _check_parquet_datasetdict(_lowerCamelCase , _lowerCamelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def _lowerCamelCase( a , a ): __a = ParquetDatasetWriter(_lowerCamelCase , tmp_path / "foo.parquet" ) assert writer.write() > 0 __a = pq.ParquetFile(tmp_path / "foo.parquet" ) __a = pf.read() assert dataset.data.table == output_table def _lowerCamelCase( a , a ): __a = str(shared_datadir / "test_image_rgb.jpg" ) __a = {"""image""": [image_path]} __a = Features({"image": Image()} ) __a = Dataset.from_dict(_lowerCamelCase , features=_lowerCamelCase ) __a = ParquetDatasetWriter(_lowerCamelCase , tmp_path / "foo.parquet" ) assert writer.write() > 0 __a = Dataset.from_parquet(str(tmp_path / "foo.parquet" ) ) assert dataset.features == reloaded_dataset.features __a = ParquetDatasetReader(str(tmp_path / "foo.parquet" ) , streaming=_lowerCamelCase ).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( "feature, expected" , [ (Features({"foo": Value("int32" )} ), None), (Features({"image": Image(), "foo": Value("int32" )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({"nested": Sequence(Audio() )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ] , ) def _lowerCamelCase( a , a ): assert get_writer_batch_size(_lowerCamelCase ) == expected
719
"""simple docstring""" from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("repo_id" , ["canonical_dataset_name", "org-name/dataset-name"] ) @pytest.mark.parametrize("path" , ["filename.csv", "filename with blanks.csv"] ) @pytest.mark.parametrize("revision" , [None, "v2"] ) def _lowerCamelCase( a , a , a ): __a = hf_hub_url(repo_id=a , path=a , revision=a ) assert url == F"https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(a )}"
67
0
"""simple docstring""" import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging SCREAMING_SNAKE_CASE__:Dict = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): _snake_case : Any = ['''input_values''', '''attention_mask'''] def __init__( self , lowerCamelCase = 1 , lowerCamelCase = 16000 , lowerCamelCase = 0.0 , lowerCamelCase = False , lowerCamelCase = 80 , lowerCamelCase = 16 , lowerCamelCase = 64 , lowerCamelCase = "hann_window" , lowerCamelCase = 1.0 , lowerCamelCase = 80 , lowerCamelCase = 7600 , lowerCamelCase = 1E-10 , lowerCamelCase = 2 , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(feature_size=A__ , sampling_rate=A__ , padding_value=A__ , **A__ ) __a = do_normalize __a = return_attention_mask __a = num_mel_bins __a = hop_length __a = win_length __a = win_function __a = frame_signal_scale __a = fmin __a = fmax __a = mel_floor __a = reduction_factor __a = win_length * sampling_rate // 1000 __a = hop_length * sampling_rate // 1000 __a = optimal_fft_length(self.sample_size ) __a = (self.n_fft // 2) + 1 __a = window_function(window_length=self.sample_size , name=self.win_function , periodic=A__ ) __a = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , A__ , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , A__ , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def a__ ( lowerCamelCase , lowerCamelCase , lowerCamelCase = 0.0 ): if attention_mask is not None: __a = np.array(A__ , np.intaa ) __a = [] for vector, length in zip(A__ , attention_mask.sum(-1 ) ): __a = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: __a = padding_value normed_input_values.append(A__ ) else: __a = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def a__ ( self , lowerCamelCase , ): __a = spectrogram( A__ , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__( self , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" F" {self.sampling_rate}. Please make sure that the provided audio input was sampled with" F" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: __a = self._process_audio( A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , **A__ , ) else: __a = None if audio_target is not None: __a = self._process_audio( A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , **A__ , ) if inputs is None: return inputs_target else: __a = inputs_target["""input_values"""] __a = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: __a = decoder_attention_mask return inputs def a__ ( self , lowerCamelCase , lowerCamelCase = False , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): __a = isinstance(A__ , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) __a = is_batched_numpy or ( isinstance(A__ , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __a = [np.asarray(A__ , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(A__ , np.ndarray ): __a = np.asarray(A__ , dtype=np.floataa ) elif isinstance(A__ , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): __a = speech.astype(np.floataa ) # always return batch if not is_batched: __a = [speech] # needed to make pad() work on spectrogram inputs __a = self.feature_size # convert into correct format for padding if is_target: __a = [self._extract_mel_features(A__ ) for waveform in speech] __a = BatchFeature({"input_values": features} ) __a = self.num_mel_bins else: __a = BatchFeature({"input_values": speech} ) __a = self.pad( A__ , padding=A__ , max_length=A__ , truncation=A__ , pad_to_multiple_of=A__ , return_attention_mask=A__ , **A__ , ) __a = feature_size_hack # convert input values to correct format __a = padded_inputs["""input_values"""] if not isinstance(input_values[0] , np.ndarray ): __a = [np.asarray(A__ , dtype=np.floataa ) for array in input_values] elif ( not isinstance(A__ , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): __a = [array.astype(np.floataa ) for array in input_values] elif isinstance(A__ , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): __a = input_values.astype(np.floataa ) # convert attention_mask to correct format __a = padded_inputs.get("attention_mask" ) if attention_mask is not None: __a = [np.asarray(A__ , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: __a = ( attention_mask if self._get_padding_strategies(A__ , max_length=A__ ) is not PaddingStrategy.DO_NOT_PAD else None ) __a = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=A__ , padding_value=self.padding_value ) if return_tensors is not None: __a = padded_inputs.convert_to_tensors(A__ ) return padded_inputs def a__ ( self ): __a = super().to_dict() # Don't serialize these as they are derived from the other properties. __a = ["""window""", """mel_filters""", """sample_size""", """sample_stride""", """n_fft""", """n_freqs"""] for name in names: if name in output: del output[name] return output
720
"""simple docstring""" from __future__ import annotations def _lowerCamelCase( a , a , a ): if len(a ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(a ) or left < -len(a ) or right >= len(a ) or right < -len(a ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] __a = (left + right) >> 1 # the middle __a = find_max(a , a , a ) # find max in range[left, mid] __a = find_max(a , mid + 1 , a ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
67
0
"""simple docstring""" def _lowerCamelCase( a ): __a = (1 + 2_4 * n) ** 0.5 return ((1 + root) / 6) % 1 == 0 def _lowerCamelCase( a = 5_0_0_0 ): __a = [(i * (3 * i - 1)) // 2 for i in range(1 , lowerCAmelCase__ )] for i, pentagonal_i in enumerate(lowerCAmelCase__ ): for j in range(lowerCAmelCase__ , len(lowerCAmelCase__ ) ): __a = pentagonal_nums[j] __a = pentagonal_i + pentagonal_j __a = pentagonal_j - pentagonal_i if is_pentagonal(lowerCAmelCase__ ) and is_pentagonal(lowerCAmelCase__ ): return b return -1 if __name__ == "__main__": print(F'''{solution() = }''')
721
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json""", """google/bigbird-roberta-large""": """https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json""", """google/bigbird-base-trivia-itc""": """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json""", # See all BigBird models at https://huggingface.co/models?filter=big_bird } class snake_case__ ( snake_case_ ): _snake_case : Any = """big_bird""" def __init__( self , lowerCamelCase=50358 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu_new" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=4096 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=True , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=66 , lowerCamelCase="block_sparse" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=64 , lowerCamelCase=3 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , sep_token_id=lowerCamelCase , **lowerCamelCase , ) __a = vocab_size __a = max_position_embeddings __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = initializer_range __a = type_vocab_size __a = layer_norm_eps __a = use_cache __a = rescale_embeddings __a = attention_type __a = use_bias __a = block_size __a = num_random_blocks __a = classifier_dropout class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task == "multiple-choice": __a = {0: "batch", 1: "choice", 2: "sequence"} else: __a = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
67
0
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Dict = { """tensor(bool)""": np.bool_, """tensor(int8)""": np.inta, """tensor(uint8)""": np.uinta, """tensor(int16)""": np.intaa, """tensor(uint16)""": np.uintaa, """tensor(int32)""": np.intaa, """tensor(uint32)""": np.uintaa, """tensor(int64)""": np.intaa, """tensor(uint64)""": np.uintaa, """tensor(float16)""": np.floataa, """tensor(float)""": np.floataa, """tensor(double)""": np.floataa, } class snake_case__ : def __init__( self , lowerCamelCase=None , **lowerCamelCase ): logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) __a = model __a = kwargs.get("model_save_dir" , UpperCamelCase__ ) __a = kwargs.get("latest_model_name" , UpperCamelCase__ ) def __call__( self , **lowerCamelCase ): __a = {k: np.array(UpperCamelCase__ ) for k, v in kwargs.items()} return self.model.run(UpperCamelCase__ , UpperCamelCase__ ) @staticmethod def a__ ( lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None ): if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) __a = "CPUExecutionProvider" return ort.InferenceSession(UpperCamelCase__ , providers=[provider] , sess_options=UpperCamelCase__ ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase ): __a = file_name if file_name is not None else ONNX_WEIGHTS_NAME __a = self.model_save_dir.joinpath(self.latest_model_name ) __a = Path(UpperCamelCase__ ).joinpath(UpperCamelCase__ ) try: shutil.copyfile(UpperCamelCase__ , UpperCamelCase__ ) except shutil.SameFileError: pass # copy external weights (for models >2GB) __a = self.model_save_dir.joinpath(UpperCamelCase__ ) if src_path.exists(): __a = Path(UpperCamelCase__ ).joinpath(UpperCamelCase__ ) try: shutil.copyfile(UpperCamelCase__ , UpperCamelCase__ ) except shutil.SameFileError: pass def a__ ( self , lowerCamelCase , **lowerCamelCase , ): if os.path.isfile(UpperCamelCase__ ): logger.error(F"Provided path ({save_directory}) should be a directory, not a file" ) return os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) # saving model weights/files self._save_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def a__ ( cls , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = False , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): __a = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(UpperCamelCase__ ): __a = OnnxRuntimeModel.load_model( os.path.join(UpperCamelCase__ , UpperCamelCase__ ) , provider=UpperCamelCase__ , sess_options=UpperCamelCase__ ) __a = Path(UpperCamelCase__ ) # load model from hub else: # download model __a = hf_hub_download( repo_id=UpperCamelCase__ , filename=UpperCamelCase__ , use_auth_token=UpperCamelCase__ , revision=UpperCamelCase__ , cache_dir=UpperCamelCase__ , force_download=UpperCamelCase__ , ) __a = Path(UpperCamelCase__ ).parent __a = Path(UpperCamelCase__ ).name __a = OnnxRuntimeModel.load_model(UpperCamelCase__ , provider=UpperCamelCase__ , sess_options=UpperCamelCase__ ) return cls(model=UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def a__ ( cls , lowerCamelCase , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , **lowerCamelCase , ): __a = None if len(str(UpperCamelCase__ ).split("@" ) ) == 2: __a , __a = model_id.split("@" ) return cls._from_pretrained( model_id=UpperCamelCase__ , revision=UpperCamelCase__ , cache_dir=UpperCamelCase__ , force_download=UpperCamelCase__ , use_auth_token=UpperCamelCase__ , **UpperCamelCase__ , )
700
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:Optional[int] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[int] = {"""tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Tuple = { """tokenizer_file""": { """bigscience/tokenizer""": """https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json""", """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json""", }, } class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = VOCAB_FILES_NAMES _snake_case : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Optional[int] = ["""input_ids""", """attention_mask"""] _snake_case : Optional[int] = None def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<unk>" , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="<pad>" , lowerCamelCase=False , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , pad_token=lowerCamelCase , add_prefix_space=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , **lowerCamelCase , ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig SCREAMING_SNAKE_CASE__:Optional[Any] = { """google/tapas-base-finetuned-sqa""": ( """https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json""" ), """google/tapas-base-finetuned-wtq""": ( """https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json""" ), """google/tapas-base-finetuned-wikisql-supervised""": ( """https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json""" ), """google/tapas-base-finetuned-tabfact""": ( """https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json""" ), } class snake_case__ ( _UpperCAmelCase ): _snake_case : Union[str, Any] = """tapas""" def __init__( self , lowerCamelCase=30522 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=1024 , lowerCamelCase=[3, 256, 256, 2, 256, 256, 10] , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=0 , lowerCamelCase=10.0 , lowerCamelCase=0 , lowerCamelCase=1.0 , lowerCamelCase=None , lowerCamelCase=1.0 , lowerCamelCase=False , lowerCamelCase=None , lowerCamelCase=1.0 , lowerCamelCase=1.0 , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase="ratio" , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=64 , lowerCamelCase=32 , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=None , lowerCamelCase=None , **lowerCamelCase , ): super().__init__(pad_token_id=lowercase__ , **lowercase__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = hidden_act __a = intermediate_size __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = type_vocab_sizes __a = initializer_range __a = layer_norm_eps # Fine-tuning task hyperparameters __a = positive_label_weight __a = num_aggregation_labels __a = aggregation_loss_weight __a = use_answer_as_supervision __a = answer_loss_importance __a = use_normalized_answer_loss __a = huber_loss_delta __a = temperature __a = aggregation_temperature __a = use_gumbel_for_cells __a = use_gumbel_for_aggregation __a = average_approximation_function __a = cell_selection_preference __a = answer_loss_cutoff __a = max_num_rows __a = max_num_columns __a = average_logits_per_cell __a = select_one_column __a = allow_empty_column_selection __a = init_cell_selection_weights_to_zero __a = reset_position_index_per_cell __a = disable_per_token_loss # Aggregation hyperparameters __a = aggregation_labels __a = no_aggregation_label_index if isinstance(self.aggregation_labels , lowercase__ ): __a = {int(lowercase__ ): v for k, v in aggregation_labels.items()}
701
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class snake_case__ : _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : int _snake_case : int _snake_case : float _snake_case : float _snake_case : Tuple[int] def a__ ( self ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def a__ ( self ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def a__ ( self ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def a__ ( self ): __a = torch.arange(self.height * self.width ) __a = torch.stack( [ pixel_indices % self.width, torch.div(lowerCamelCase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def a__ ( self ): __a , *__a = self.shape __a = int(np.prod(lowerCamelCase ) ) __a = self.get_image_coords() __a = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __a = self.get_camera_rays(lowerCamelCase ) __a = rays.view(lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def a__ ( self , lowerCamelCase ): __a , *__a , __a = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __a = coords.view(lowerCamelCase , -1 , 2 ) __a = self.resolution() __a = self.fov() __a = (flat.float() / (res - 1)) * 2 - 1 __a = fracs * torch.tan(fov / 2 ) __a = fracs.view(lowerCamelCase , -1 , 2 ) __a = ( self.z.view(lowerCamelCase , 1 , 3 ) + self.x.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) __a = directions / directions.norm(dim=-1 , keepdim=lowerCamelCase ) __a = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCamelCase , *lowerCamelCase , 2 , 3 ) def a__ ( self , lowerCamelCase , lowerCamelCase ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCamelCase , height=lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowerCamelCase( a ): __a = [] __a = [] __a = [] __a = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __a = np.array([np.sin(a ), np.cos(a ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __a = -z * 4 __a = np.array([np.cos(a ), -np.sin(a ), 0.0] ) __a = np.cross(a , a ) origins.append(a ) xs.append(a ) ys.append(a ) zs.append(a ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(a , axis=0 ) ).float() , x=torch.from_numpy(np.stack(a , axis=0 ) ).float() , y=torch.from_numpy(np.stack(a , axis=0 ) ).float() , z=torch.from_numpy(np.stack(a , axis=0 ) ).float() , width=a , height=a , x_fov=0.7 , y_fov=0.7 , shape=(1, len(a )) , )
67
0
"""simple docstring""" import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class snake_case__ : def __init__( self , lowerCamelCase = "cpu" , lowerCamelCase = "openai/clip-vit-large-patch14" ): __a = device __a = CLIPTokenizerFast.from_pretrained(lowerCamelCase ) __a = [0.4814_5466, 0.457_8275, 0.4082_1073] __a = [0.2686_2954, 0.2613_0258, 0.2757_7711] __a = torchvision.transforms.Normalize(self.image_mean , self.image_std ) __a = torchvision.transforms.Resize(224 ) __a = torchvision.transforms.CenterCrop(224 ) def a__ ( self , lowerCamelCase ): __a = self.resize(lowerCamelCase ) __a = self.center_crop(lowerCamelCase ) __a = self.normalize(lowerCamelCase ) return images def __call__( self , lowerCamelCase=None , lowerCamelCase=None , **lowerCamelCase ): __a = self.tokenizer(text=lowerCamelCase , **lowerCamelCase ) __a = self.preprocess_img(lowerCamelCase ) __a = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class snake_case__ ( nn.Module ): def __init__( self , lowerCamelCase=10 , lowerCamelCase=0.01 , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase="image" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=False , ): super().__init__() __a = None __a = device if device else get_device() if vqgan: __a = vqgan else: __a = load_vqgan(self.device , conf_path=lowerCamelCase , ckpt_path=lowerCamelCase ) self.vqgan.eval() if clip: __a = clip else: __a = CLIPModel.from_pretrained("openai/clip-vit-base-patch32" ) self.clip.to(self.device ) __a = ProcessorGradientFlow(device=self.device ) __a = iterations __a = lr __a = log __a = make_grid __a = return_val __a = quantize __a = self.vqgan.decoder.z_shape def a__ ( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=5 , lowerCamelCase=True ): __a = [] if output_path is None: __a = './animation.gif' if input_path is None: __a = self.save_path __a = sorted(glob(input_path + "/*" ) ) if not len(lowerCamelCase ): raise ValueError( "No images found in save path, aborting (did you pass save_intermediate=True to the generate" " function?)" ) if len(lowerCamelCase ) == 1: print("Only one image found in save path, (did you pass save_intermediate=True to the generate function?)" ) __a = total_duration / len(lowerCamelCase ) __a = [frame_duration] * len(lowerCamelCase ) if extend_frames: __a = 1.5 __a = 3 for file_name in paths: if file_name.endswith(".png" ): images.append(imageio.imread(lowerCamelCase ) ) imageio.mimsave(lowerCamelCase , lowerCamelCase , duration=lowerCamelCase ) print(F"gif saved to {output_path}" ) def a__ ( self , lowerCamelCase=None , lowerCamelCase=None ): if not (path or img): raise ValueError("Input either path or tensor" ) if img is not None: raise NotImplementedError __a = preprocess(Image.open(lowerCamelCase ) , target_image_size=256 ).to(self.device ) __a = preprocess_vqgan(lowerCamelCase ) __a = self.vqgan.encode(lowerCamelCase ) return z def a__ ( self , lowerCamelCase ): __a = self.latent.detach().requires_grad_() __a = base_latent + transform_vector if self.quantize: __a = self.vqgan.quantize(lowerCamelCase ) else: __a = trans_latent return self.vqgan.decode(lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=None ): __a = self.clip_preprocessor(text=lowerCamelCase , images=lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ) __a = self.clip(**lowerCamelCase ) __a = clip_outputs.logits_per_image if weights is not None: __a = similarity_logits * weights return similarity_logits.sum() def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self._get_clip_similarity(pos_prompts["prompts"] , lowerCamelCase , weights=(1 / pos_prompts["weights"]) ) if neg_prompts: __a = self._get_clip_similarity(neg_prompts["prompts"] , lowerCamelCase , weights=neg_prompts["weights"] ) else: __a = torch.tensor([1] , device=self.device ) __a = -torch.log(lowerCamelCase ) + torch.log(lowerCamelCase ) return loss def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = torch.randn_like(self.latent , requires_grad=lowerCamelCase , device=self.device ) __a = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() __a = self._add_vector(lowerCamelCase ) __a = loop_post_process(lowerCamelCase ) __a = self._get_CLIP_loss(lowerCamelCase , lowerCamelCase , lowerCamelCase ) print("CLIP loss" , lowerCamelCase ) if self.log: wandb.log({"CLIP Loss": clip_loss} ) clip_loss.backward(retain_graph=lowerCamelCase ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): wandb.init(reinit=lowerCamelCase , project="face-editor" ) wandb.config.update({"Positive Prompts": positive_prompts} ) wandb.config.update({"Negative Prompts": negative_prompts} ) wandb.config.update({"lr": self.lr, "iterations": self.iterations} ) if image_path: __a = Image.open(lowerCamelCase ) __a = image.resize((256, 256) ) wandb.log("Original Image" , wandb.Image(lowerCamelCase ) ) def a__ ( self , lowerCamelCase ): if not prompts: return [] __a = [] __a = [] if isinstance(lowerCamelCase , lowerCamelCase ): __a = [prompt.strip() for prompt in prompts.split("|" )] for prompt in prompts: if isinstance(lowerCamelCase , (tuple, list) ): __a = prompt[0] __a = float(prompt[1] ) elif ":" in prompt: __a = prompt.split(":" ) __a = float(lowerCamelCase ) else: __a = prompt __a = 1.0 processed_prompts.append(lowerCamelCase ) weights.append(lowerCamelCase ) return { "prompts": processed_prompts, "weights": torch.tensor(lowerCamelCase , device=self.device ), } def a__ ( self , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=None , ): if image_path: __a = self._get_latent(lowerCamelCase ) else: __a = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(lowerCamelCase , lowerCamelCase , lowerCamelCase ) assert pos_prompts, "You must provide at least one positive prompt." __a = self.process_prompts(lowerCamelCase ) __a = self.process_prompts(lowerCamelCase ) if save_final and save_path is None: __a = os.path.join("./outputs/" , "_".join(pos_prompts["prompts"] ) ) if not os.path.exists(lowerCamelCase ): os.makedirs(lowerCamelCase ) else: __a = save_path + '_' + get_timestamp() os.makedirs(lowerCamelCase ) __a = save_path __a = self.vqgan.decode(self.latent )[0] if show_intermediate: print("Original Image" ) show_pil(custom_to_pil(lowerCamelCase ) ) __a = loop_post_process(lowerCamelCase ) for iter, transformed_img in enumerate(self._optimize_CLIP(lowerCamelCase , lowerCamelCase , lowerCamelCase ) ): if show_intermediate: show_pil(lowerCamelCase ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}.png" ) ) if self.log: wandb.log({"Image": wandb.Image(lowerCamelCase )} ) if show_final: show_pil(lowerCamelCase ) if save_final: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}_final.png" ) )
702
"""simple docstring""" def _lowerCamelCase( a ): return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase( a ): __a = 0 __a = number while duplicate > 0: __a , __a = divmod(a , 1_0 ) fact_sum += factorial(a ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__:Optional[Any] = int(input("""Enter number: """).strip()) print( F'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
67
0
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class snake_case__ ( __A, __A, unittest.TestCase ): _snake_case : Any = IFInpaintingPipeline _snake_case : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} _snake_case : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _snake_case : Union[str, Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def a__ ( self ): return self._get_dummy_components() def a__ ( self , lowerCamelCase , lowerCamelCase=0 ): if str(lowerCamelCase ).startswith("mps" ): __a = torch.manual_seed(lowerCamelCase ) else: __a = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __a = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __a = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __a = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def a__ ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def a__ ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def a__ ( self ): super().test_save_load_floataa(expected_max_diff=1E-1 ) def a__ ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def a__ ( self ): self._test_save_load_local() def a__ ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
703
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE__:Optional[Any] = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Union[str, Any] = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class snake_case__ ( metaclass=__lowercase ): _snake_case : Any = ["""keras_nlp"""] def __init__( self , *lowerCamelCase , **lowerCamelCase ): requires_backends(self , ["keras_nlp"] )
704
"""simple docstring""" import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def _lowerCamelCase( a , a , a ): __a = OmegaConf.load(a ) __a = torch.load(a , map_location="cpu" )["model"] __a = list(state_dict.keys() ) # extract state_dict for VQVAE __a = {} __a = "first_stage_model." for key in keys: if key.startswith(a ): __a = state_dict[key] # extract state_dict for UNetLDM __a = {} __a = "model.diffusion_model." for key in keys: if key.startswith(a ): __a = state_dict[key] __a = config.model.params.first_stage_config.params __a = config.model.params.unet_config.params __a = VQModel(**a ).eval() vqvae.load_state_dict(a ) __a = UNetLDMModel(**a ).eval() unet.load_state_dict(a ) __a = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=a , ) __a = LDMPipeline(a , a , a ) pipeline.save_pretrained(a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
67
0
"""simple docstring""" from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def _lowerCamelCase( a , a , a , a ): for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F"Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})" else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F"Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})" def _lowerCamelCase( a , a , a , a , a=True ): model.train() __a = model(lowerCAmelCase__ ) __a = F.mse_loss(lowerCAmelCase__ , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(lowerCAmelCase__ ) def _lowerCamelCase( a , a=False ): set_seed(4_2 ) __a = RegressionModel() __a = deepcopy(lowerCAmelCase__ ) __a = RegressionDataset(length=8_0 ) __a = DataLoader(lowerCAmelCase__ , batch_size=1_6 ) model.to(accelerator.device ) if sched: __a = AdamW(params=model.parameters() , lr=1E-3 ) __a = AdamW(params=ddp_model.parameters() , lr=1E-3 ) __a = LambdaLR(lowerCAmelCase__ , lr_lambda=lambda a : epoch**0.65 ) __a = LambdaLR(lowerCAmelCase__ , lr_lambda=lambda a : epoch**0.65 ) # Make a copy of `model` if sched: __a , __a , __a , __a = accelerator.prepare(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: __a , __a = accelerator.prepare(lowerCAmelCase__ , lowerCAmelCase__ ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def _lowerCamelCase( a ): # Test when on a single CPU or GPU that the context manager does nothing __a , __a , __a = get_training_setup(lowerCAmelCase__ ) # Use a single batch __a , __a = next(iter(lowerCAmelCase__ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model __a , __a = accelerator.gather((ddp_input, ddp_target) ) __a , __a = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCAmelCase__ ): step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: # Sync grads step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F"Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})" # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) __a = ddp_input[torch.randperm(len(lowerCAmelCase__ ) )] def _lowerCamelCase( a ): # Test on distributed setup that context manager behaves properly __a , __a , __a = get_training_setup(lowerCAmelCase__ ) # Use a single batch __a , __a = next(iter(lowerCAmelCase__ ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model __a , __a = accelerator.gather((ddp_input, ddp_target) ) __a , __a = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(lowerCAmelCase__ ): step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: # Sync grads step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F"Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})" else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F"Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})" # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) __a = ddp_input[torch.randperm(len(lowerCAmelCase__ ) )] def _lowerCamelCase( a=False , a=False ): __a = Accelerator( split_batches=lowerCAmelCase__ , dispatch_batches=lowerCAmelCase__ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly __a , __a , __a = get_training_setup(lowerCAmelCase__ ) for iteration, batch in enumerate(lowerCAmelCase__ ): __a , __a = batch.values() # Gather the distributed inputs and targs for the base model __a , __a = accelerator.gather((ddp_input, ddp_target) ) __a , __a = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Do "gradient accumulation" (noop) with accelerator.accumulate(lowerCAmelCase__ ): step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(lowerCAmelCase__ ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F"Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})" else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F"Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})" # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) __a = ddp_input[torch.randperm(len(lowerCAmelCase__ ) )] GradientState._reset_state() def _lowerCamelCase( a=False , a=False ): __a = Accelerator( split_batches=lowerCAmelCase__ , dispatch_batches=lowerCAmelCase__ , gradient_accumulation_steps=2 ) # Test that context manager behaves properly __a , __a , __a , __a , __a , __a , __a = get_training_setup(lowerCAmelCase__ , lowerCAmelCase__ ) for iteration, batch in enumerate(lowerCAmelCase__ ): __a , __a = batch.values() # Gather the distributed inputs and targs for the base model __a , __a = accelerator.gather((ddp_input, ddp_target) ) __a , __a = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(lowerCAmelCase__ )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(lowerCAmelCase__ ): step_model(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F"Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n" __a = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(lowerCAmelCase__ )) if accelerator.num_processes > 1: check_model_parameters(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Shuffle ddp_input on each iteration torch.manual_seed(1_3_3_7 + iteration ) GradientState._reset_state() def _lowerCamelCase( ): __a = Accelerator() __a = RegressionDataset(length=8_0 ) __a = DataLoader(lowerCAmelCase__ , batch_size=1_6 ) __a = RegressionDataset(length=9_6 ) __a = DataLoader(lowerCAmelCase__ , batch_size=1_6 ) __a , __a = accelerator.prepare(lowerCAmelCase__ , lowerCAmelCase__ ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(lowerCAmelCase__ ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCAmelCase__ ) if iteration < len(lowerCAmelCase__ ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(lowerCAmelCase__ ): assert id(accelerator.gradient_state.active_dataloader ) == id(lowerCAmelCase__ ) if batch_num < len(lowerCAmelCase__ ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def _lowerCamelCase( ): __a = Accelerator() __a = accelerator.state if state.local_process_index == 0: print("**Test `accumulate` gradient accumulation with dataloader break**" ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print("**Test NOOP `no_sync` context manager**" ) test_noop_sync(lowerCAmelCase__ ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print("**Test Distributed `no_sync` context manager**" ) test_distributed_sync(lowerCAmelCase__ ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( "**Test `accumulate` gradient accumulation, " , F"`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**" , ) test_gradient_accumulation(lowerCAmelCase__ , lowerCAmelCase__ ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version("<" , "2.0" ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( "**Test `accumulate` gradient accumulation with optimizer and scheduler, " , "`split_batches=False`, `dispatch_batches=False`**" , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( "**Test `accumulate` gradient accumulation with optimizer and scheduler, " , F"`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**" , ) test_gradient_accumulation_with_opt_and_scheduler(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowerCamelCase( a ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
705
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[Any] = { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json""", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case__ ( snake_case_ ): _snake_case : str = """blenderbot-small""" _snake_case : str = ["""past_key_values"""] _snake_case : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCamelCase=50265 , lowerCamelCase=512 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="gelu" , lowerCamelCase=512 , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , lowerCamelCase=1 , lowerCamelCase=False , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=2 , **lowerCamelCase , ): __a = vocab_size __a = max_position_embeddings __a = d_model __a = encoder_ffn_dim __a = encoder_layers __a = encoder_attention_heads __a = decoder_ffn_dim __a = decoder_layers __a = decoder_attention_heads __a = dropout __a = attention_dropout __a = activation_dropout __a = activation_function __a = init_std __a = encoder_layerdrop __a = decoder_layerdrop __a = use_cache __a = encoder_layers __a = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , is_encoder_decoder=lowerCamelCase , decoder_start_token_id=lowerCamelCase , forced_eos_token_id=lowerCamelCase , **lowerCamelCase , ) class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a = {0: "batch"} __a = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __a = {0: "batch", 1: "decoder_sequence"} __a = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} else: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = super().outputs else: __a = super(lowerCamelCase , self ).outputs if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Generate decoder inputs __a = seq_length if not self.use_past else 1 __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) __a = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} __a = dict(**lowerCamelCase , **lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape __a = common_inputs["decoder_input_ids"].shape[1] __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = decoder_seq_length + 3 __a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __a = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase )] , dim=1 ) __a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __a , __a = self.num_layers __a = min(lowerCamelCase , lowerCamelCase ) __a = max(lowerCamelCase , lowerCamelCase ) - min_num_layers __a = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), ) ) # TODO: test this. __a = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(lowerCamelCase , lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape # Not using the same length for past_key_values __a = seqlen + 2 __a , __a = self.num_layers __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = common_inputs["attention_mask"].dtype __a = torch.cat( [common_inputs["attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) __a = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(lowerCamelCase ) ] return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __a = tokenizer.num_special_tokens_to_add(lowerCamelCase ) __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence __a = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __a = dict(tokenizer(lowerCamelCase , return_tensors=lowerCamelCase ) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): if self.task in ["default", "seq2seq-lm"]: __a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) elif self.task == "causal-lm": __a = self._generate_dummy_inputs_for_causal_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) else: __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if self.task in ["default", "seq2seq-lm"]: __a = super()._flatten_past_key_values_(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __a = super(lowerCamelCase , self )._flatten_past_key_values_( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import numpy as np def _lowerCamelCase( a ): return 1 / (1 + np.exp(-vector )) def _lowerCamelCase( a ): return vector * sigmoid(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
706
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=99 , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=9 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase=8 , lowerCamelCase=0.1 , lowerCamelCase=0.002 , lowerCamelCase=1 , lowerCamelCase=0 , lowerCamelCase=0 , lowerCamelCase=None , lowerCamelCase=None , ): __a = parent __a = batch_size __a = encoder_seq_length __a = decoder_seq_length # For common tests __a = self.decoder_seq_length __a = is_training __a = use_attention_mask __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = d_ff __a = relative_attention_num_buckets __a = dropout_rate __a = initializer_factor __a = eos_token_id __a = pad_token_id __a = decoder_start_token_id __a = None __a = decoder_layers def a__ ( self ): return TaConfig.from_pretrained("google/umt5-base" ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): if attention_mask is None: __a = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __a = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __a = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=lowerCamelCase ) if decoder_head_mask is None: __a = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) if cross_attn_head_mask is None: __a = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def a__ ( self ): __a = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __a = input_ids.clamp(self.pad_token_id + 1 ) __a = decoder_input_ids.clamp(self.pad_token_id + 1 ) __a = self.get_config() __a = config.num_attention_heads __a = self.prepare_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, input_dict def a__ ( self ): __a , __a = self.prepare_config_and_inputs() return config, inputs_dict def a__ ( self ): return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self ): return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase , attention_mask=lowerCamelCase , decoder_attention_mask=lowerCamelCase , ) __a = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ) __a = result.last_hidden_state __a = result.past_key_values __a = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).get_decoder().to(lowerCamelCase ).eval() # first forward pass __a = model(lowerCamelCase , use_cache=lowerCamelCase ) __a = model(lowerCamelCase ) __a = model(lowerCamelCase , use_cache=lowerCamelCase ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) + 1 ) __a , __a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = model(lowerCamelCase )["last_hidden_state"] __a = model(lowerCamelCase , past_key_values=lowerCamelCase )["last_hidden_state"] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -1, random_slice_idx].detach() __a = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).to(lowerCamelCase ).half().eval() __a = model(**lowerCamelCase )["last_hidden_state"] self.parent.assertFalse(torch.isnan(lowerCamelCase ).any().item() ) @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _snake_case : int = (UMTaForConditionalGeneration,) if is_torch_available() else () _snake_case : Optional[int] = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _snake_case : List[Any] = True _snake_case : Union[str, Any] = False _snake_case : Union[str, Any] = False _snake_case : Tuple = True _snake_case : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests _snake_case : Optional[Any] = [0.8, 0.9] def a__ ( self ): __a = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() __a = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=lowerCamelCase , opset_version=9 , input_names=["input_ids", "decoder_input_ids"] , ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase ) def a__ ( self ): __a = ["encoder_attentions", "decoder_attentions", "cross_attentions"] __a = self.model_tester.prepare_config_and_inputs() __a = config_and_inputs[0] __a = UMTaForConditionalGeneration(lowerCamelCase ).eval() model.to(lowerCamelCase ) __a = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=lowerCamelCase ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), } for attn_name, (name, mask) in zip(lowerCamelCase , head_masking.items() ): __a = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": __a = torch.ones( config.num_decoder_layers , config.num_heads , device=lowerCamelCase ) __a = model.generate( config_and_inputs[1]["input_ids"] , num_beams=1 , max_length=3 , output_attentions=lowerCamelCase , return_dict_in_generate=lowerCamelCase , **lowerCamelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step __a = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def a__ ( self ): pass @require_torch @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def a__ ( self ): __a = UMTaForConditionalGeneration.from_pretrained("google/umt5-small" , return_dict=lowerCamelCase ).to(lowerCamelCase ) __a = AutoTokenizer.from_pretrained("google/umt5-small" , use_fast=lowerCamelCase , legacy=lowerCamelCase ) __a = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] __a = tokenizer(lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ).input_ids # fmt: off __a = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase , lowerCamelCase ) __a = model.generate(input_ids.to(lowerCamelCase ) ) __a = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] __a = tokenizer.batch_decode(lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=18 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , ): __a = size if size is not None else {"height": 18, "width": 18} __a = parent __a = batch_size __a = num_channels __a = image_size __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize def a__ ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8866_4436_3403_3203, 0.6618_8293_6954_4983, 0.3891_7464_0178_6804], [-0.6042_5591_4688_1104, -0.0_2295_0088_6052_8469, 0.5423_7973_6900_3296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : List[str] = ImageGPTImageProcessor if is_vision_available() else None def a__ ( self ): __a = ImageGPTImageProcessingTester(self ) @property def a__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "clusters" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) def a__ ( self ): __a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) __a = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) __a = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(lowerCamelCase , obj[key] ) ) else: self.assertEqual(obj[key] , lowerCamelCase ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __a = os.path.join(lowerCamelCase , "image_processor.json" ) image_processor_first.to_json_file(lowerCamelCase ) __a = self.image_processing_class.from_json_file(lowerCamelCase ).to_dict() __a = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(lowerCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , lowerCamelCase ) def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(lowerCamelCase ) __a = self.image_processing_class.from_pretrained(lowerCamelCase ).to_dict() __a = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(lowerCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , lowerCamelCase ) @unittest.skip("ImageGPT requires clusters at initialization" ) def a__ ( self ): pass def _lowerCamelCase( ): __a = load_dataset("hf-internal-testing/fixtures_image_utils" , split="test" ) __a = Image.open(dataset[4]["file"] ) __a = Image.open(dataset[5]["file"] ) __a = [imagea, imagea] return images @require_vision @require_torch class snake_case__ ( unittest.TestCase ): @slow def a__ ( self ): __a = ImageGPTImageProcessor.from_pretrained("openai/imagegpt-small" ) __a = prepare_images() # test non-batched __a = image_processing(images[0] , return_tensors="pt" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1024) ) __a = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , lowerCamelCase ) # test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1024) ) __a = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , lowerCamelCase )
707
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase( a , a , a ): # Initialise PyTorch model __a = MobileBertConfig.from_json_file(a ) print(F"Building PyTorch model from configuration: {config}" ) __a = MobileBertForPreTraining(a ) # Load weights from tf checkpoint __a = load_tf_weights_in_mobilebert(a , a , a ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) SCREAMING_SNAKE_CASE__:List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
67
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE__:Tuple = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:List[str] = { 'salesforce/blip2-opt-2.7b': 'https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json', } class snake_case__ ( lowerCAmelCase__ ): _snake_case : Optional[int] = "blip_2_vision_model" def __init__( self , lowerCamelCase=1408 , lowerCamelCase=6144 , lowerCamelCase=39 , lowerCamelCase=16 , lowerCamelCase=224 , lowerCamelCase=14 , lowerCamelCase="gelu" , lowerCamelCase=0.0_0001 , lowerCamelCase=0.0 , lowerCamelCase=1E-10 , lowerCamelCase=True , **lowerCamelCase , ): super().__init__(**_SCREAMING_SNAKE_CASE ) __a = hidden_size __a = intermediate_size __a = num_hidden_layers __a = num_attention_heads __a = patch_size __a = image_size __a = initializer_range __a = attention_dropout __a = layer_norm_eps __a = hidden_act __a = qkv_bias @classmethod def a__ ( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE ) __a , __a = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": __a = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) class snake_case__ ( lowerCAmelCase__ ): _snake_case : str = "blip_2_qformer" def __init__( self , lowerCamelCase=30522 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=512 , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=2 , lowerCamelCase=1408 , **lowerCamelCase , ): super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = hidden_act __a = intermediate_size __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = initializer_range __a = layer_norm_eps __a = position_embedding_type __a = cross_attention_frequency __a = encoder_hidden_size @classmethod def a__ ( cls , lowerCamelCase , **lowerCamelCase ): cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE ) __a , __a = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("model_type" ) == "blip-2": __a = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) class snake_case__ ( lowerCAmelCase__ ): _snake_case : Any = "blip-2" _snake_case : Optional[Any] = True def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=32 , **lowerCamelCase ): super().__init__(**_SCREAMING_SNAKE_CASE ) if vision_config is None: __a = {} logger.info("vision_config is None. initializing the Blip2VisionConfig with default values." ) if qformer_config is None: __a = {} logger.info("qformer_config is None. Initializing the Blip2QFormerConfig with default values." ) if text_config is None: __a = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) __a = BlipaVisionConfig(**_SCREAMING_SNAKE_CASE ) __a = BlipaQFormerConfig(**_SCREAMING_SNAKE_CASE ) __a = text_config["model_type"] if "model_type" in text_config else "opt" __a = CONFIG_MAPPING[text_model_type](**_SCREAMING_SNAKE_CASE ) __a = self.text_config.tie_word_embeddings __a = self.text_config.is_encoder_decoder __a = num_query_tokens __a = self.vision_config.hidden_size __a = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __a = 1.0 __a = 0.02 @classmethod def a__ ( cls , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **_SCREAMING_SNAKE_CASE , ) def a__ ( self ): __a = copy.deepcopy(self.__dict__ ) __a = self.vision_config.to_dict() __a = self.qformer_config.to_dict() __a = self.text_config.to_dict() __a = self.__class__.model_type return output
708
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__ ( snake_case_ ): def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __a = input_file.read() __a = regexp.search(lowerCamelCase ) return match def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __a = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __a = regexp.finditer(lowerCamelCase ) __a = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCamelCase ) ): raise AssertionError(F"open(...) must use utf-8 encoding in {dataset}" ) def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCamelCase ) ): raise AssertionError(F"print statement found in {dataset}. Use datasets.logger/logging instead." )
67
0
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_A ) class snake_case__ ( _A ): _snake_case : Dict = field(default="""text-classification""", metadata={"""include_in_asdict_even_if_is_default""": True} ) _snake_case : int = Features({"""text""": Value("""string""" )} ) _snake_case : int = Features({"""labels""": ClassLabel} ) _snake_case : List[str] = """text""" _snake_case : Optional[int] = """labels""" def a__ ( self , lowerCamelCase ): if self.label_column not in features: raise ValueError(F"Column {self.label_column} is not present in features." ) if not isinstance(features[self.label_column] , UpperCamelCase__ ): raise ValueError(F"Column {self.label_column} is not a ClassLabel." ) __a = copy.deepcopy(self ) __a = self.label_schema.copy() __a = features[self.label_column] __a = label_schema return task_template @property def a__ ( self ): return { self.text_column: "text", self.label_column: "labels", }
709
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
67
0
"""simple docstring""" import torch from diffusers import DiffusionPipeline class snake_case__ ( snake_case__ ): def __init__( self , lowerCamelCase , lowerCamelCase ): super().__init__() self.register_modules(unet=_A , scheduler=_A ) def __call__( self ): __a = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) __a = 1 __a = self.unet(_A , _A ).sample __a = self.scheduler.step(_A , _A , _A ).prev_sample __a = scheduler_output - scheduler_output + torch.ones_like(_A ) return result
710
"""simple docstring""" import heapq import sys import numpy as np SCREAMING_SNAKE_CASE__:Optional[int] = tuple[int, int] class snake_case__ : def __init__( self ): __a = [] __a = set() def a__ ( self ): if not self.empty(): return self.elements[0][0] else: return float("inf" ) def a__ ( self ): return len(self.elements ) == 0 def a__ ( self , lowerCamelCase , lowerCamelCase ): if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(lowerCamelCase ) else: # update # print("update", item) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def a__ ( self , lowerCamelCase ): if item in self.set: self.set.remove(lowerCamelCase ) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def a__ ( self ): return self.elements[0][1] def a__ ( self ): ((__a) , (__a)) = heapq.heappop(self.elements ) self.set.remove(lowerCamelCase ) return (priority, item) def _lowerCamelCase( a , a ): # euclidean distance __a = np.array(a ) __a = np.array(a ) return np.linalg.norm(a - b ) def _lowerCamelCase( a , a ): # integer division by time variable return consistent_heuristic(a , a ) // t def _lowerCamelCase( a , a ): # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _lowerCamelCase( a , a , a , a ): __a = g_function[start] + Wa * heuristics[i](a , a ) return ans def _lowerCamelCase( a , a , a ): __a = np.chararray((n, n) ) for i in range(a ): for j in range(a ): __a = "*" for i in range(a ): for j in range(a ): if (j, (n - 1) - i) in blocks: __a = "#" __a = "-" __a = back_pointer[goal] while x != start: ((__a) , (__a)) = x # print(x) __a = "-" __a = back_pointer[x] __a = "-" for i in range(a ): for j in range(a ): if (i, j) == (0, n - 1): print(grid[i][j] , end=" " ) print("<-- End position" , end=" " ) else: print(grid[i][j] , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) print("PATH TAKEN BY THE ALGORITHM IS:-" ) __a = back_pointer[goal] while x != start: print(a , end=" " ) __a = back_pointer[x] print(a ) sys.exit() def _lowerCamelCase( a ): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _lowerCamelCase( a , a , a , a , a , a , a , a , ): for itera in range(a ): open_list[itera].remove_element(a ) # print("s", s) # print("j", j) ((__a) , (__a)) = s __a = (x - 1, y) __a = (x + 1, y) __a = (x, y + 1) __a = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(a ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(a ) __a = -1 __a = float("inf" ) if valid(a ) and g_function[neighbours] > g_function[s] + 1: __a = g_function[s] + 1 __a = s if neighbours not in close_list_anchor: open_list[0].put(a , key(a , 0 , a , a ) ) if neighbours not in close_list_inad: for var in range(1 , a ): if key(a , a , a , a ) <= Wa * key( a , 0 , a , a ): open_list[j].put( a , key(a , a , a , a ) ) def _lowerCamelCase( ): __a = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(1_5 , 2_0 ): some_list.append((x, 1_7) ) for x in range(1_0 , 1_9 ): for y in range(1 , 1_5 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(1_2 , 1_9 ): some_list.append((x, y) ) for x in range(3 , 1_3 ): for y in range(1_6 , 1_9 ): some_list.append((x, y) ) return some_list SCREAMING_SNAKE_CASE__:Any = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} SCREAMING_SNAKE_CASE__:str = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] SCREAMING_SNAKE_CASE__:int = make_common_ground() SCREAMING_SNAKE_CASE__:List[str] = blocks_blk # hyper parameters SCREAMING_SNAKE_CASE__:str = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 20 SCREAMING_SNAKE_CASE__:Dict = 3 # one consistent and two other inconsistent # start and end destination SCREAMING_SNAKE_CASE__:Dict = (0, 0) SCREAMING_SNAKE_CASE__:Optional[Any] = (n - 1, n - 1) SCREAMING_SNAKE_CASE__:List[str] = 1 def _lowerCamelCase( a , a , a ): __a = {start: 0, goal: float("inf" )} __a = {start: -1, goal: -1} __a = [] __a = set() for i in range(a ): open_list.append(PriorityQueue() ) open_list[i].put(a , key(a , a , a , a ) ) __a = [] __a = [] while open_list[0].minkey() < float("inf" ): for i in range(1 , a ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a , __a = open_list[i].top_show() visited.add(a ) expand_state( a , a , a , a , a , a , a , a , ) close_list_inad.append(a ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a = open_list[0].top_show() visited.add(a ) expand_state( a , 0 , a , a , a , a , a , a , ) close_list_anchor.append(a ) print("No path found to goal" ) print() for i in range(n - 1 , -1 , -1 ): for j in range(a ): if (j, i) in blocks: print("#" , end=" " ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" " ) else: print("-" , end=" " ) else: print("*" , end=" " ) if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
67
0
"""simple docstring""" from __future__ import annotations def _lowerCamelCase( a ): if not nums: raise ValueError("List is empty" ) return sum(__snake_case ) / len(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
711
"""simple docstring""" SCREAMING_SNAKE_CASE__:Any = """Alexander Joslin""" import operator as op from .stack import Stack def _lowerCamelCase( a ): __a = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} __a = Stack() __a = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a ) ) elif i in operators: # RULE 2 operator_stack.push(a ) elif i == ")": # RULE 4 __a = operator_stack.peek() operator_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operators[opr](a , a ) operand_stack.push(a ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
67
0
"""simple docstring""" import functools from typing import Any def _lowerCamelCase( a , a ): if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or len(lowerCamelCase_ ) == 0: raise ValueError("the string should be not empty string" ) if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or not all( isinstance(lowerCamelCase_ , lowerCamelCase_ ) and len(lowerCamelCase_ ) > 0 for item in words ): raise ValueError("the words should be a list of non-empty strings" ) # Build trie __a = {} __a = """WORD_KEEPER""" for word in words: __a = trie for c in word: if c not in trie_node: __a = {} __a = trie_node[c] __a = True __a = len(lowerCamelCase_ ) # Dynamic programming method @functools.cache def is_breakable(a ) -> bool: if index == len_string: return True __a = trie for i in range(lowerCamelCase_ , lowerCamelCase_ ): __a = trie_node.get(string[i] , lowerCamelCase_ ) if trie_node is None: return False if trie_node.get(lowerCamelCase_ , lowerCamelCase_ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
712
"""simple docstring""" from math import pi def _lowerCamelCase( a , a ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
67
0
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) SCREAMING_SNAKE_CASE__:Optional[Any] = """hf-internal-testing/tiny-random-bert""" SCREAMING_SNAKE_CASE__:int = os.path.join(TRANSFORMERS_CACHE, """models--hf-internal-testing--tiny-random-bert""") SCREAMING_SNAKE_CASE__:List[str] = """9b8c223d42b2188cb49d29af482996f9d0f3e5a6""" class snake_case__ ( unittest.TestCase ): def a__ ( self ): __a = cached_file(UpperCamelCase_ , UpperCamelCase_ ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(UpperCamelCase_ ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) ) ) with open(os.path.join(UpperCamelCase_ , "refs" , "main" ) ) as f: __a = f.read() self.assertEqual(UpperCamelCase_ , os.path.join(UpperCamelCase_ , "snapshots" , UpperCamelCase_ , UpperCamelCase_ ) ) self.assertTrue(os.path.isfile(UpperCamelCase_ ) ) # File is cached at the same place the second time. __a = cached_file(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(UpperCamelCase_ , UpperCamelCase_ ) # Using a specific revision to test the full commit hash. __a = cached_file(UpperCamelCase_ , UpperCamelCase_ , revision="9b8c223" ) self.assertEqual(UpperCamelCase_ , os.path.join(UpperCamelCase_ , "snapshots" , UpperCamelCase_ , UpperCamelCase_ ) ) def a__ ( self ): with self.assertRaisesRegex(UpperCamelCase_ , "is not a valid model identifier" ): __a = cached_file("tiny-random-bert" , UpperCamelCase_ ) with self.assertRaisesRegex(UpperCamelCase_ , "is not a valid git identifier" ): __a = cached_file(UpperCamelCase_ , UpperCamelCase_ , revision="aaaa" ) with self.assertRaisesRegex(UpperCamelCase_ , "does not appear to have a file named" ): __a = cached_file(UpperCamelCase_ , "conf" ) def a__ ( self ): with self.assertRaisesRegex(UpperCamelCase_ , "does not appear to have a file named" ): __a = cached_file(UpperCamelCase_ , "conf" ) with open(os.path.join(UpperCamelCase_ , "refs" , "main" ) ) as f: __a = f.read() self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase_ , ".no_exist" , UpperCamelCase_ , "conf" ) ) ) __a = cached_file(UpperCamelCase_ , "conf" , _raise_exceptions_for_missing_entries=UpperCamelCase_ ) self.assertIsNone(UpperCamelCase_ ) __a = cached_file(UpperCamelCase_ , "conf" , local_files_only=UpperCamelCase_ , _raise_exceptions_for_missing_entries=UpperCamelCase_ ) self.assertIsNone(UpperCamelCase_ ) __a = mock.Mock() __a = 500 __a = {} __a = HTTPError __a = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=UpperCamelCase_ ) as mock_head: __a = cached_file(UpperCamelCase_ , "conf" , _raise_exceptions_for_connection_errors=UpperCamelCase_ ) self.assertIsNone(UpperCamelCase_ ) # This check we did call the fake head request mock_head.assert_called() def a__ ( self ): self.assertTrue(has_file("hf-internal-testing/tiny-bert-pt-only" , UpperCamelCase_ ) ) self.assertFalse(has_file("hf-internal-testing/tiny-bert-pt-only" , UpperCamelCase_ ) ) self.assertFalse(has_file("hf-internal-testing/tiny-bert-pt-only" , UpperCamelCase_ ) ) def a__ ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo("bert-base-cased" , "ahah.txt" ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(UpperCamelCase_ , "is not a valid model identifier" ): get_file_from_repo("bert-base-case" , UpperCamelCase_ ) # The function raises if the revision does not exist. with self.assertRaisesRegex(UpperCamelCase_ , "is not a valid git identifier" ): get_file_from_repo("bert-base-cased" , UpperCamelCase_ , revision="ahaha" ) __a = get_file_from_repo("bert-base-cased" , UpperCamelCase_ ) # The name is the cached name which is not very easy to test, so instead we load the content. __a = json.loads(open(UpperCamelCase_ , "r" ).read() ) self.assertEqual(config["hidden_size"] , 768 ) def a__ ( self ): with tempfile.TemporaryDirectory() as tmp_dir: __a = Path(UpperCamelCase_ ) / "a.txt" filename.touch() self.assertEqual(get_file_from_repo(UpperCamelCase_ , "a.txt" ) , str(UpperCamelCase_ ) ) self.assertIsNone(get_file_from_repo(UpperCamelCase_ , "b.txt" ) )
713
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): _snake_case : Dict = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = IMAGENET_DEFAULT_MEAN , lowerCamelCase = IMAGENET_DEFAULT_STD , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __a = image_std if image_std is not None else IMAGENET_DEFAULT_STD def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __a = int((256 / 224) * size["shortest_edge"] ) __a = get_resize_output_image_size(lowerCamelCase , size=lowerCamelCase , default_to_square=lowerCamelCase ) __a = {"height": output_size[0], "width": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( F"Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}" ) return resize( lowerCamelCase , size=(size_dict["height"], size_dict["width"]) , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"Size dict must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" ) __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(lowerCamelCase , lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(lowerCamelCase , lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(lowerCamelCase , lowerCamelCase , lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
"""simple docstring""" def _lowerCamelCase( a , a ): if digit_amount > 0: return round(number - int(lowerCamelCase__ ) , lowerCamelCase__ ) return number - int(lowerCamelCase__ ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
714
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=30 , lowerCamelCase=2 , lowerCamelCase=3 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=10 , lowerCamelCase=0.02 , lowerCamelCase=None , lowerCamelCase=2 , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def a__ ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, pixel_values, labels def a__ ( self ): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = ViTForMaskedImageModeling(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __a = 1 __a = ViTForMaskedImageModeling(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = self.type_sequence_label_size __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __a = 1 __a = ViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def a__ ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, unittest.TestCase ): _snake_case : Any = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) _snake_case : List[Any] = ( {"""feature-extraction""": ViTModel, """image-classification""": ViTForImageClassification} if is_torch_available() else {} ) _snake_case : int = True _snake_case : int = False _snake_case : str = False _snake_case : Optional[Any] = False def a__ ( self ): __a = ViTModelTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def a__ ( self ): pass def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def a__ ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(lowerCamelCase ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def a__ ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def _lowerCamelCase( ): __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def a__ ( self ): return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def a__ ( self ): __a = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(lowerCamelCase ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(**lowerCamelCase ) # verify the logits __a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __a = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def a__ ( self ): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. __a = ViTModel.from_pretrained("facebook/dino-vits8" ).to(lowerCamelCase ) __a = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass with torch.no_grad(): __a = model(lowerCamelCase , interpolate_pos_encoding=lowerCamelCase ) # verify the logits __a = torch.Size((1, 3601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , lowerCamelCase ) __a = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def a__ ( self ): __a = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) __a = self.default_image_processor __a = prepare_img() __a = image_processor(images=lowerCamelCase , return_tensors="pt" ) __a = inputs.pixel_values.to(lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __a = model(lowerCamelCase )
67
0
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE__:List[Any] = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.multihead_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def _lowerCamelCase( a , a , a ): __a = state_dict.pop(a__ ) __a = val def _lowerCamelCase( a ): __a = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: __a = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) __a = value else: __a = value return new_state_dict def _lowerCamelCase( a ): __a = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) __a = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight" ) __a = state_dict.pop(F"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict __a = in_proj_weight[:2_5_6, :] __a = in_proj_bias[:2_5_6] __a = in_proj_weight[2_5_6:5_1_2, :] __a = in_proj_bias[2_5_6:5_1_2] __a = in_proj_weight[-2_5_6:, :] __a = in_proj_bias[-2_5_6:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention __a = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight" ) __a = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias" ) # next, add query, keys and values (in that order) to the state dict __a = in_proj_weight[:2_5_6, :] __a = in_proj_bias[:2_5_6] __a = in_proj_weight[2_5_6:5_1_2, :] __a = in_proj_bias[2_5_6:5_1_2] __a = in_proj_weight[-2_5_6:, :] __a = in_proj_bias[-2_5_6:] # read in weights + bias of input projection layer of cross-attention __a = state_dict.pop( F"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight" ) __a = state_dict.pop(F"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias" ) # next, add query, keys and values (in that order) of cross-attention to the state dict __a = in_proj_weight_cross_attn[:2_5_6, :] __a = in_proj_bias_cross_attn[:2_5_6] __a = in_proj_weight_cross_attn[2_5_6:5_1_2, :] __a = in_proj_bias_cross_attn[2_5_6:5_1_2] __a = in_proj_weight_cross_attn[-2_5_6:, :] __a = in_proj_bias_cross_attn[-2_5_6:] def _lowerCamelCase( a , a ): __a , __a = image.size __a = max(a__ , a__ ) __a = 8_0_0 if "detection" in checkpoint_url else 1_0_0_0 __a = target_max_size / current_max_size __a = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def _lowerCamelCase( a ): __a = F.to_tensor(a__ ) __a = F.normalize(a__ , mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ) return image @torch.no_grad() def _lowerCamelCase( a , a , a ): logger.info("Converting model..." ) # load original state dict __a = torch.hub.load_state_dict_from_url(a__ , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(a__ , a__ , a__ ) __a = rename_backbone_keys(a__ ) # query, key and value matrices need special treatment read_in_q_k_v(a__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them __a = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): __a = state_dict.pop(a__ ) __a = val # create HuggingFace model and load state dict __a = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: __a = 1_5 __a = 2 __a = {0: "table", 1: "table rotated"} __a = idalabel __a = {v: k for k, v in idalabel.items()} else: __a = 1_2_5 __a = 6 __a = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } __a = idalabel __a = {v: k for k, v in idalabel.items()} __a = DetrImageProcessor( format="coco_detection" , max_size=8_0_0 if "detection" in checkpoint_url else 1_0_0_0 ) __a = TableTransformerForObjectDetection(a__ ) model.load_state_dict(a__ ) model.eval() # verify our conversion __a = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" __a = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=a__ ) __a = Image.open(a__ ).convert("RGB" ) __a = normalize(resize(a__ , a__ ) ).unsqueeze(0 ) __a = model(a__ ) if "detection" in checkpoint_url: __a = (1, 1_5, 3) __a = torch.tensor( [[-6.78_97, -16.99_85, 6.79_37], [-8.01_86, -22.21_92, 6.96_77], [-7.31_17, -21.07_08, 7.40_55]] ) __a = torch.tensor([[0.48_67, 0.17_67, 0.67_32], [0.67_18, 0.44_79, 0.38_30], [0.47_16, 0.17_60, 0.63_64]] ) else: __a = (1, 1_2_5, 7) __a = torch.tensor( [[-18.14_30, -8.32_14, 4.82_74], [-18.46_85, -7.13_61, -4.26_67], [-26.36_93, -9.34_29, -4.99_62]] ) __a = torch.tensor([[0.49_83, 0.55_95, 0.94_40], [0.49_16, 0.63_15, 0.59_54], [0.61_08, 0.86_37, 0.11_35]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , a__ , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , a__ , atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(a__ ).mkdir(exist_ok=a__ ) model.save_pretrained(a__ ) image_processor.save_pretrained(a__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) __a = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(a__ ) image_processor.push_to_hub(a__ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
715
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCamelCase , lowerCamelCase=7 , lowerCamelCase=3 , lowerCamelCase=30 , lowerCamelCase=400 , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=[0.5, 0.5, 0.5] , lowerCamelCase=True , lowerCamelCase=1 / 255 , lowerCamelCase=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __a = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __a = parent __a = batch_size __a = num_channels __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std __a = do_rescale __a = rescale_factor __a = do_pad def a__ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def a__ ( self , lowerCamelCase , lowerCamelCase=False ): if not batched: __a = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __a , __a = image.size else: __a , __a = image.shape[1], image.shape[2] if w < h: __a = int(self.size["shortest_edge"] * h / w ) __a = self.size["shortest_edge"] elif w > h: __a = self.size["shortest_edge"] __a = int(self.size["shortest_edge"] * w / h ) else: __a = self.size["shortest_edge"] __a = self.size["shortest_edge"] else: __a = [] for image in image_inputs: __a , __a = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __a = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class snake_case__ ( snake_case_, unittest.TestCase ): _snake_case : List[Any] = DetaImageProcessor if is_vision_available() else None def a__ ( self ): __a = DetaImageProcessingTester(self ) @property def a__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def a__ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_rescale" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_pad" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def a__ ( self ): __a = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def a__ ( self ): pass def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def a__ ( self ): # Initialize image_processing __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __a = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __a = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __a , __a = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def a__ ( self ): # prepare image and target __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"image_id": 39769, "annotations": target} # encode them __a = DetaImageProcessor() __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def a__ ( self ): # prepare image, target and masks_path __a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __a = json.loads(f.read() ) __a = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __a = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __a = DetaImageProcessor(format="coco_panoptic" ) __a = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __a = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __a = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __a = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __a = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __a = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __a = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __a = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __a = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __a = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __a = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __a = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
67
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor SCREAMING_SNAKE_CASE__:Any = logging.get_logger(__name__) class UpperCamelCase ( __lowerCAmelCase ): def __init__( self , *lowerCamelCase , **lowerCamelCase ): warnings.warn( "The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ChineseCLIPImageProcessor instead." , lowerCAmelCase_ , ) super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ )
716
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE__:Dict = logging.getLogger() def _lowerCamelCase( ): __a = argparse.ArgumentParser() parser.add_argument("-f" ) __a = parser.parse_args() return args.f class snake_case__ ( snake_case_ ): def a__ ( self ): __a = logging.StreamHandler(sys.stdout ) logger.addHandler(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowerCamelCase , "argv" , lowerCamelCase ): __a = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowerCamelCase , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self ): __a = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase ) __a = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowerCamelCase )
67
0
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 SCREAMING_SNAKE_CASE__:Optional[Any] = sys.version_info >= (3, 10) def _lowerCamelCase( a=None , a=None ): return field(default_factory=lambda: default , metadata=a ) @dataclass class snake_case__ : _snake_case : int _snake_case : float _snake_case : str _snake_case : bool @dataclass class snake_case__ : _snake_case : int = 42 _snake_case : str = field(default="""toto""", metadata={"""help""": """help message"""} ) @dataclass class snake_case__ : _snake_case : bool = False _snake_case : bool = True _snake_case : Optional[bool] = None class snake_case__ ( snake_case_ ): _snake_case : Tuple = 'titi' _snake_case : Tuple = 'toto' class snake_case__ ( snake_case_ ): _snake_case : int = 'titi' _snake_case : Dict = 'toto' _snake_case : List[Any] = 42 @dataclass class snake_case__ : _snake_case : BasicEnum = "toto" def a__ ( self ): __a = BasicEnum(self.foo ) @dataclass class snake_case__ : _snake_case : MixedTypeEnum = "toto" def a__ ( self ): __a = MixedTypeEnum(self.foo ) @dataclass class snake_case__ : _snake_case : Optional[int] = None _snake_case : Optional[float] = field(default=snake_case_, metadata={"""help""": """help message"""} ) _snake_case : Optional[str] = None _snake_case : Optional[List[str]] = list_field(default=[] ) _snake_case : Optional[List[int]] = list_field(default=[] ) @dataclass class snake_case__ : _snake_case : List[int] = list_field(default=[] ) _snake_case : List[int] = list_field(default=[1, 2, 3] ) _snake_case : List[str] = list_field(default=["""Hallo""", """Bonjour""", """Hello"""] ) _snake_case : List[float] = list_field(default=[0.1, 0.2, 0.3] ) @dataclass class snake_case__ : _snake_case : List[int] = field() _snake_case : str = field() _snake_case : BasicEnum = field() def a__ ( self ): __a = BasicEnum(self.required_enum ) @dataclass class snake_case__ : _snake_case : int _snake_case : "BasicEnum" = field() _snake_case : "Optional[bool]" = None _snake_case : "str" = field(default="""toto""", metadata={"""help""": """help message"""} ) _snake_case : "List[str]" = list_field(default=["""Hallo""", """Bonjour""", """Hello"""] ) if is_python_no_less_than_3_10: @dataclass class snake_case__ : _snake_case : bool = False _snake_case : bool = True _snake_case : bool | None = None @dataclass class snake_case__ : _snake_case : int | None = None _snake_case : float | None = field(default=snake_case_, metadata={"""help""": """help message"""} ) _snake_case : str | None = None _snake_case : list[str] | None = list_field(default=[] ) _snake_case : list[int] | None = list_field(default=[] ) class snake_case__ ( unittest.TestCase ): def a__ ( self , lowerCamelCase , lowerCamelCase ): self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): __a = {k: v for k, v in vars(UpperCAmelCase__ ).items() if k != '''container'''} __a = {k: v for k, v in vars(UpperCAmelCase__ ).items() if k != '''container'''} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("choices" , UpperCAmelCase__ ) and yy.get("choices" , UpperCAmelCase__ ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["type"](UpperCAmelCase__ ) , yy["type"](UpperCAmelCase__ ) ) del xx["type"], yy["type"] self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument("--foo" , type=UpperCAmelCase__ , required=UpperCAmelCase__ ) expected.add_argument("--bar" , type=UpperCAmelCase__ , required=UpperCAmelCase__ ) expected.add_argument("--baz" , type=UpperCAmelCase__ , required=UpperCAmelCase__ ) expected.add_argument("--flag" , type=UpperCAmelCase__ , default=UpperCAmelCase__ , const=UpperCAmelCase__ , nargs="?" ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __a = ['''--foo''', '''1''', '''--baz''', '''quux''', '''--bar''', '''0.5'''] (__a ) = parser.parse_args_into_dataclasses(UpperCAmelCase__ , look_for_args_file=UpperCAmelCase__ ) self.assertFalse(example.flag ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument("--foo" , default=42 , type=UpperCAmelCase__ ) expected.add_argument("--baz" , default="toto" , type=UpperCAmelCase__ , help="help message" ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = argparse.ArgumentParser() expected.add_argument("--foo" , type=UpperCAmelCase__ , default=UpperCAmelCase__ , const=UpperCAmelCase__ , nargs="?" ) expected.add_argument("--baz" , type=UpperCAmelCase__ , default=UpperCAmelCase__ , const=UpperCAmelCase__ , nargs="?" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("--no_baz" , action="store_false" , default=UpperCAmelCase__ , dest="baz" ) expected.add_argument("--opt" , type=UpperCAmelCase__ , default=UpperCAmelCase__ ) __a = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(UpperCAmelCase__ ) for dataclass_type in dataclass_types: __a = HfArgumentParser(UpperCAmelCase__ ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_args([] ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=UpperCAmelCase__ , baz=UpperCAmelCase__ , opt=UpperCAmelCase__ ) ) __a = parser.parse_args(["--foo", "--no_baz"] ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=UpperCAmelCase__ , baz=UpperCAmelCase__ , opt=UpperCAmelCase__ ) ) __a = parser.parse_args(["--foo", "--baz"] ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=UpperCAmelCase__ , baz=UpperCAmelCase__ , opt=UpperCAmelCase__ ) ) __a = parser.parse_args(["--foo", "True", "--baz", "True", "--opt", "True"] ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=UpperCAmelCase__ , baz=UpperCAmelCase__ , opt=UpperCAmelCase__ ) ) __a = parser.parse_args(["--foo", "False", "--baz", "False", "--opt", "False"] ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=UpperCAmelCase__ , baz=UpperCAmelCase__ , opt=UpperCAmelCase__ ) ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=["titi", "toto", 42] , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) __a = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) __a = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) __a = parser.parse_args_into_dataclasses(["--foo", "titi"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) __a = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) __a = parser.parse_args_into_dataclasses(["--foo", "42"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def a__ ( self ): @dataclass class snake_case__ : _snake_case : Literal["titi", "toto", 42] = "toto" __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=("titi", "toto", 42) , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) __a = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) __a = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument("--foo_int" , nargs="+" , default=[] , type=UpperCAmelCase__ ) expected.add_argument("--bar_int" , nargs="+" , default=[1, 2, 3] , type=UpperCAmelCase__ ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=UpperCAmelCase__ ) expected.add_argument("--foo_float" , nargs="+" , default=[0.1, 0.2, 0.3] , type=UpperCAmelCase__ ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_args([] ) self.assertEqual( UpperCAmelCase__ , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["Hallo", "Bonjour", "Hello"] , foo_float=[0.1, 0.2, 0.3] ) , ) __a = parser.parse_args("--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7".split() ) self.assertEqual(UpperCAmelCase__ , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["a", "b", "c"] , foo_float=[0.1, 0.7] ) ) def a__ ( self ): __a = argparse.ArgumentParser() expected.add_argument("--foo" , default=UpperCAmelCase__ , type=UpperCAmelCase__ ) expected.add_argument("--bar" , default=UpperCAmelCase__ , type=UpperCAmelCase__ , help="help message" ) expected.add_argument("--baz" , default=UpperCAmelCase__ , type=UpperCAmelCase__ ) expected.add_argument("--ces" , nargs="+" , default=[] , type=UpperCAmelCase__ ) expected.add_argument("--des" , nargs="+" , default=[] , type=UpperCAmelCase__ ) __a = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(UpperCAmelCase__ ) for dataclass_type in dataclass_types: __a = HfArgumentParser(UpperCAmelCase__ ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_args([] ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=UpperCAmelCase__ , bar=UpperCAmelCase__ , baz=UpperCAmelCase__ , ces=[] , des=[] ) ) __a = parser.parse_args("--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3".split() ) self.assertEqual(UpperCAmelCase__ , Namespace(foo=12 , bar=3.14 , baz="42" , ces=["a", "b", "c"] , des=[1, 2, 3] ) ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument("--required_list" , nargs="+" , type=UpperCAmelCase__ , required=UpperCAmelCase__ ) expected.add_argument("--required_str" , type=UpperCAmelCase__ , required=UpperCAmelCase__ ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=UpperCAmelCase__ , ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = argparse.ArgumentParser() expected.add_argument("--foo" , type=UpperCAmelCase__ , required=UpperCAmelCase__ ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=UpperCAmelCase__ , ) expected.add_argument("--opt" , type=UpperCAmelCase__ , default=UpperCAmelCase__ ) expected.add_argument("--baz" , default="toto" , type=UpperCAmelCase__ , help="help message" ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=UpperCAmelCase__ ) self.argparsersEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = { '''foo''': 12, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, } __a = parser.parse_dict(UpperCAmelCase__ )[0] __a = BasicExample(**UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = { '''foo''': 12, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, '''extra''': 42, } self.assertRaises(UpperCAmelCase__ , parser.parse_dict , UpperCAmelCase__ , allow_extra_keys=UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = { '''foo''': 12, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, } with tempfile.TemporaryDirectory() as tmp_dir: __a = os.path.join(UpperCAmelCase__ , "temp_json" ) os.mkdir(UpperCAmelCase__ ) with open(temp_local_path + ".json" , "w+" ) as f: json.dump(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_yaml_file(Path(temp_local_path + ".json" ) )[0] __a = BasicExample(**UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) __a = { '''foo''': 12, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, } with tempfile.TemporaryDirectory() as tmp_dir: __a = os.path.join(UpperCAmelCase__ , "temp_yaml" ) os.mkdir(UpperCAmelCase__ ) with open(temp_local_path + ".yaml" , "w+" ) as f: yaml.dump(UpperCAmelCase__ , UpperCAmelCase__ ) __a = parser.parse_yaml_file(Path(temp_local_path + ".yaml" ) )[0] __a = BasicExample(**UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def a__ ( self ): __a = HfArgumentParser(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ )
717
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging SCREAMING_SNAKE_CASE__:Union[str, Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = ["""pixel_values"""] def __init__( self , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = True , lowerCamelCase = 1 / 255 , lowerCamelCase = True , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = True , **lowerCamelCase , ): super().__init__(**lowerCamelCase ) __a = size if size is not None else {"shortest_edge": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) __a = crop_size if crop_size is not None else {"height": 224, "width": 224} __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase , param_name="crop_size" ) __a = do_resize __a = size __a = resample __a = do_center_crop __a = crop_size __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __a = image_std if image_std is not None else OPENAI_CLIP_STD __a = do_convert_rgb def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = PILImageResampling.BICUBIC , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase , default_to_square=lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) __a = get_resize_output_image_size(lowerCamelCase , size=size["shortest_edge"] , default_to_square=lowerCamelCase ) return resize(lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): __a = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowerCamelCase , size=(size["height"], size["width"]) , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase = None , **lowerCamelCase , ): return normalize(lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = None , lowerCamelCase = ChannelDimension.FIRST , **lowerCamelCase , ): __a = do_resize if do_resize is not None else self.do_resize __a = size if size is not None else self.size __a = get_size_dict(lowerCamelCase , param_name="size" , default_to_square=lowerCamelCase ) __a = resample if resample is not None else self.resample __a = do_center_crop if do_center_crop is not None else self.do_center_crop __a = crop_size if crop_size is not None else self.crop_size __a = get_size_dict(lowerCamelCase , param_name="crop_size" , default_to_square=lowerCamelCase ) __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __a = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __a = [convert_to_rgb(lowerCamelCase ) for image in images] # All transformations expect numpy arrays. __a = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __a = [self.resize(image=lowerCamelCase , size=lowerCamelCase , resample=lowerCamelCase ) for image in images] if do_center_crop: __a = [self.center_crop(image=lowerCamelCase , size=lowerCamelCase ) for image in images] if do_rescale: __a = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_normalize: __a = [self.normalize(image=lowerCamelCase , mean=lowerCamelCase , std=lowerCamelCase ) for image in images] __a = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __a = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
67
0
from statistics import mean import numpy as np def _lowerCamelCase( a , a , a , a ): __a = 0 # Number of processes finished __a = 0 # Displays the finished process. # If it is 0, the performance is completed if it is 1, before the performance. __a = [0] * no_of_process # List to include calculation results __a = [0] * no_of_process # Sort by arrival time. __a = [burst_time[i] for i in np.argsort(a )] __a = [process_name[i] for i in np.argsort(a )] arrival_time.sort() while no_of_process > finished_process_count: __a = 0 while finished_process[i] == 1: i += 1 if current_time < arrival_time[i]: __a = arrival_time[i] __a = 0 # Index showing the location of the process being performed __a = 0 # Saves the current response ratio. __a = 0 for i in range(0 , a ): if finished_process[i] == 0 and arrival_time[i] <= current_time: __a = (burst_time[i] + (current_time - arrival_time[i])) / burst_time[ i ] if response_ratio < temp: __a = temp __a = i # Calculate the turn around time __a = current_time + burst_time[loc] - arrival_time[loc] current_time += burst_time[loc] # Indicates that the process has been performed. __a = 1 # Increase finished_process_count by 1 finished_process_count += 1 return turn_around_time def _lowerCamelCase( a , a , a , a ): __a = [0] * no_of_process for i in range(0 , a ): __a = turn_around_time[i] - burst_time[i] return waiting_time if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Optional[int] = 5 SCREAMING_SNAKE_CASE__:Optional[int] = ["""A""", """B""", """C""", """D""", """E"""] SCREAMING_SNAKE_CASE__:Union[str, Any] = [1, 2, 3, 4, 5] SCREAMING_SNAKE_CASE__:Tuple = [1, 2, 3, 4, 5] SCREAMING_SNAKE_CASE__:Union[str, Any] = calculate_turn_around_time( process_name, arrival_time, burst_time, no_of_process ) SCREAMING_SNAKE_CASE__:int = calculate_waiting_time( process_name, turn_around_time, burst_time, no_of_process ) print("""Process name \tArrival time \tBurst time \tTurn around time \tWaiting time""") for i in range(0, no_of_process): print( F'''{process_name[i]}\t\t{arrival_time[i]}\t\t{burst_time[i]}\t\t''' F'''{turn_around_time[i]}\t\t\t{waiting_time[i]}''' ) print(F'''average waiting time : {mean(waiting_time):.5f}''') print(F'''average turn around time : {mean(turn_around_time):.5f}''')
718
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Any = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Optional[Any] = { """vocab_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/vocab.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/vocab.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/vocab.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/vocab.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/vocab.json""", }, """merges_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/merges.txt""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/merges.txt""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/merges.txt""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/merges.txt""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/merges.txt""", }, """tokenizer_file""": { """gpt2""": """https://huggingface.co/gpt2/resolve/main/tokenizer.json""", """gpt2-medium""": """https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json""", """gpt2-large""": """https://huggingface.co/gpt2-large/resolve/main/tokenizer.json""", """gpt2-xl""": """https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json""", """distilgpt2""": """https://huggingface.co/distilgpt2/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__:Union[str, Any] = { """gpt2""": 1024, """gpt2-medium""": 1024, """gpt2-large""": 1024, """gpt2-xl""": 1024, """distilgpt2""": 1024, } class snake_case__ ( snake_case_ ): _snake_case : Tuple = VOCAB_FILES_NAMES _snake_case : str = PRETRAINED_VOCAB_FILES_MAP _snake_case : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : List[str] = ["""input_ids""", """attention_mask"""] _snake_case : Dict = GPTaTokenizer def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase="<|endoftext|>" , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , add_prefix_space=lowerCamelCase , **lowerCamelCase , ) __a = kwargs.pop("add_bos_token" , lowerCamelCase ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
from math import isqrt def _lowerCamelCase( a ): return all(number % divisor != 0 for divisor in range(2 , isqrt(__a ) + 1 ) ) def _lowerCamelCase( a = 1_0**6 ): __a = 0 __a = 1 __a = 7 while prime_candidate < max_prime: primes_count += is_prime(__a ) cube_index += 1 prime_candidate += 6 * cube_index return primes_count if __name__ == "__main__": print(F'''{solution() = }''')
719
"""simple docstring""" from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("repo_id" , ["canonical_dataset_name", "org-name/dataset-name"] ) @pytest.mark.parametrize("path" , ["filename.csv", "filename with blanks.csv"] ) @pytest.mark.parametrize("revision" , [None, "v2"] ) def _lowerCamelCase( a , a , a ): __a = hf_hub_url(repo_id=a , path=a , revision=a ) assert url == F"https://huggingface.co/datasets/{repo_id}/resolve/{revision or 'main'}/{quote(a )}"
67
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE__:Any = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Dict = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
720
"""simple docstring""" from __future__ import annotations def _lowerCamelCase( a , a , a ): if len(a ) == 0: raise ValueError("find_max() arg is an empty sequence" ) if ( left >= len(a ) or left < -len(a ) or right >= len(a ) or right < -len(a ) ): raise IndexError("list index out of range" ) if left == right: return nums[left] __a = (left + right) >> 1 # the middle __a = find_max(a , a , a ) # find max in range[left, mid] __a = find_max(a , mid + 1 , a ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
67
0
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__:Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Dict = {"""vocab_file""": """spiece.model"""} SCREAMING_SNAKE_CASE__:Optional[int] = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", } } SCREAMING_SNAKE_CASE__:Tuple = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } SCREAMING_SNAKE_CASE__:Any = """▁""" class snake_case__ ( lowercase__ ): _snake_case : Tuple = VOCAB_FILES_NAMES _snake_case : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowerCamelCase , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase="[CLS]" , lowerCamelCase="[SEP]" , lowerCamelCase="<unk>" , lowerCamelCase="[SEP]" , lowerCamelCase="<pad>" , lowerCamelCase="[CLS]" , lowerCamelCase="[MASK]" , lowerCamelCase = None , **lowerCamelCase , ): __a = ( AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ , normalized=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token ) __a = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , remove_space=UpperCAmelCase__ , keep_accents=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) __a = do_lower_case __a = remove_space __a = keep_accents __a = vocab_file __a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) @property def a__ ( self ): return len(self.sp_model ) def a__ ( self ): __a = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __a = self.__dict__.copy() __a = None return state def __setstate__( self , lowerCamelCase ): __a = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __a = {} __a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def a__ ( self , lowerCamelCase ): if self.remove_space: __a = ''' '''.join(inputs.strip().split() ) else: __a = inputs __a = outputs.replace("``" , "\"" ).replace("\'\'" , "\"" ) if not self.keep_accents: __a = unicodedata.normalize("NFKD" , UpperCAmelCase__ ) __a = ''''''.join([c for c in outputs if not unicodedata.combining(UpperCAmelCase__ )] ) if self.do_lower_case: __a = outputs.lower() return outputs def a__ ( self , lowerCamelCase ): __a = self.preprocess_text(UpperCAmelCase__ ) __a = self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) __a = [] for piece in pieces: if len(UpperCAmelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): __a = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: __a = cur_pieces[1:] else: __a = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase__ ) else: new_pieces.append(UpperCAmelCase__ ) return new_pieces def a__ ( self , lowerCamelCase ): return self.sp_model.PieceToId(UpperCAmelCase__ ) def a__ ( self , lowerCamelCase ): return self.sp_model.IdToPiece(UpperCAmelCase__ ) def a__ ( self , lowerCamelCase ): __a = [] __a = '''''' __a = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(UpperCAmelCase__ ) + token __a = True __a = [] else: current_sub_tokens.append(UpperCAmelCase__ ) __a = False out_string += self.sp_model.decode(UpperCAmelCase__ ) return out_string.strip() def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is not None: return [1] + ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1] return [1] + ([0] * len(UpperCAmelCase__ )) + [1] def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if not os.path.isdir(UpperCAmelCase__ ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __a = os.path.join( UpperCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , "wb" ) as fi: __a = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,)
721
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = { """google/bigbird-roberta-base""": """https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json""", """google/bigbird-roberta-large""": """https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json""", """google/bigbird-base-trivia-itc""": """https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json""", # See all BigBird models at https://huggingface.co/models?filter=big_bird } class snake_case__ ( snake_case_ ): _snake_case : Any = """big_bird""" def __init__( self , lowerCamelCase=50358 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu_new" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=4096 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=True , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=66 , lowerCamelCase="block_sparse" , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=64 , lowerCamelCase=3 , lowerCamelCase=None , **lowerCamelCase , ): super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , sep_token_id=lowerCamelCase , **lowerCamelCase , ) __a = vocab_size __a = max_position_embeddings __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = initializer_range __a = type_vocab_size __a = layer_norm_eps __a = use_cache __a = rescale_embeddings __a = attention_type __a = use_bias __a = block_size __a = num_random_blocks __a = classifier_dropout class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task == "multiple-choice": __a = {0: "batch", 1: "choice", 2: "sequence"} else: __a = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
67
0
"""simple docstring""" import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class snake_case__ ( unittest.TestCase ): _snake_case : List[Any] = inspect.getfile(accelerate.test_utils ) _snake_case : str = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_cli.py"""] ) _snake_case : Union[str, Any] = ["accelerate", "launch"] _snake_case : Tuple = Path.home() / ".cache/huggingface/accelerate" _snake_case : Tuple = "default_config.yaml" _snake_case : Dict = config_folder / config_file _snake_case : List[str] = config_folder / "_default_config.yaml" _snake_case : Tuple = Path("""tests/test_configs""" ) @classmethod def a__ ( cls ): if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def a__ ( cls ): if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def a__ ( self ): __a = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] , env=os.environ.copy() ) def a__ ( self ): for config in sorted(self.test_config_path.glob("**/*.yaml" ) ): with self.subTest(config_file=_lowerCamelCase ): execute_subprocess_async( self.base_cmd + ["--config_file", str(_lowerCamelCase ), self.test_file_path] , env=os.environ.copy() ) def a__ ( self ): execute_subprocess_async(["accelerate", "test"] , env=os.environ.copy() ) class snake_case__ ( unittest.TestCase ): _snake_case : Any = "test-tpu" _snake_case : Optional[Any] = "us-central1-a" _snake_case : Dict = "ls" _snake_case : List[str] = ["accelerate", "tpu-config"] _snake_case : Optional[int] = "cd /usr/share" _snake_case : Union[str, Any] = "tests/test_samples/test_command_file.sh" _snake_case : Any = "Running gcloud compute tpus tpu-vm ssh" def a__ ( self ): __a = run_command( self.cmd + ["--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug"] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--debug"] , return_stdout=_lowerCamelCase ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--debug"] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--command", "echo \"Hello World\"", "--debug", ] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command_file", self.command_file, "--debug"] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command_file", self.command_file, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--debug"] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all" , _lowerCamelCase , ) def a__ ( self ): __a = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--accelerate_version", "12.0.0", "--debug", ] , return_stdout=_lowerCamelCase , ) self.assertIn( F"{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all" , _lowerCamelCase , )
700
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__:Optional[int] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[int] = {"""tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE__:Tuple = { """tokenizer_file""": { """bigscience/tokenizer""": """https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json""", """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json""", }, } class snake_case__ ( snake_case_ ): _snake_case : Optional[Any] = VOCAB_FILES_NAMES _snake_case : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : Optional[int] = ["""input_ids""", """attention_mask"""] _snake_case : Optional[int] = None def __init__( self , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase="<unk>" , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="<pad>" , lowerCamelCase=False , lowerCamelCase=False , **lowerCamelCase , ): super().__init__( lowerCamelCase , lowerCamelCase , tokenizer_file=lowerCamelCase , unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , pad_token=lowerCamelCase , add_prefix_space=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , **lowerCamelCase , ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , lowerCamelCase ) != add_prefix_space: __a = getattr(lowerCamelCase , pre_tok_state.pop("type" ) ) __a = add_prefix_space __a = pre_tok_class(**lowerCamelCase ) __a = add_prefix_space def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , *lowerCamelCase , **lowerCamelCase ): __a = kwargs.get("is_split_into_words" , lowerCamelCase ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with" " pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase , **lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = None ): __a = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) + [self.eos_token_id] ) if len(lowerCamelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
67
0
"""simple docstring""" def _lowerCamelCase( a ): if n == 1 or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): return 0 elif n == 2: return 1 else: __a = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def _lowerCamelCase( a ): __a = 0 __a = 2 while digits < n: index += 1 __a = len(str(fibonacci(lowerCamelCase_ ) ) ) return index def _lowerCamelCase( a = 1_0_0_0 ): return fibonacci_digits_index(lowerCamelCase_ ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
701
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class snake_case__ : _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : torch.Tensor # [batch_size x 3] _snake_case : int _snake_case : int _snake_case : float _snake_case : float _snake_case : Tuple[int] def a__ ( self ): assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def a__ ( self ): return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def a__ ( self ): return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def a__ ( self ): __a = torch.arange(self.height * self.width ) __a = torch.stack( [ pixel_indices % self.width, torch.div(lowerCamelCase , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def a__ ( self ): __a , *__a = self.shape __a = int(np.prod(lowerCamelCase ) ) __a = self.get_image_coords() __a = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __a = self.get_camera_rays(lowerCamelCase ) __a = rays.view(lowerCamelCase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def a__ ( self , lowerCamelCase ): __a , *__a , __a = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __a = coords.view(lowerCamelCase , -1 , 2 ) __a = self.resolution() __a = self.fov() __a = (flat.float() / (res - 1)) * 2 - 1 __a = fracs * torch.tan(fov / 2 ) __a = fracs.view(lowerCamelCase , -1 , 2 ) __a = ( self.z.view(lowerCamelCase , 1 , 3 ) + self.x.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(lowerCamelCase , 1 , 3 ) * fracs[:, :, 1:] ) __a = directions / directions.norm(dim=-1 , keepdim=lowerCamelCase ) __a = torch.stack( [ torch.broadcast_to(self.origin.view(lowerCamelCase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(lowerCamelCase , *lowerCamelCase , 2 , 3 ) def a__ ( self , lowerCamelCase , lowerCamelCase ): assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=lowerCamelCase , height=lowerCamelCase , x_fov=self.x_fov , y_fov=self.y_fov , ) def _lowerCamelCase( a ): __a = [] __a = [] __a = [] __a = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __a = np.array([np.sin(a ), np.cos(a ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __a = -z * 4 __a = np.array([np.cos(a ), -np.sin(a ), 0.0] ) __a = np.cross(a , a ) origins.append(a ) xs.append(a ) ys.append(a ) zs.append(a ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(a , axis=0 ) ).float() , x=torch.from_numpy(np.stack(a , axis=0 ) ).float() , y=torch.from_numpy(np.stack(a , axis=0 ) ).float() , z=torch.from_numpy(np.stack(a , axis=0 ) ).float() , width=a , height=a , x_fov=0.7 , y_fov=0.7 , shape=(1, len(a )) , )
67
0
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar SCREAMING_SNAKE_CASE__:str = TypeVar("""_T""") class snake_case__ ( Generic[_T] ): def __init__( self , lowerCamelCase = None ): __a = list(iterable or [] ) __a = [] def __len__( self ): return len(self._stacka ) + len(self._stacka ) def __repr__( self ): return F"Queue({tuple(self._stacka[::-1] + self._stacka )})" def a__ ( self , lowerCamelCase ): self._stacka.append(_lowercase ) def a__ ( self ): __a = self._stacka.pop __a = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("Queue is empty" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
702
"""simple docstring""" def _lowerCamelCase( a ): return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def _lowerCamelCase( a ): __a = 0 __a = number while duplicate > 0: __a , __a = divmod(a , 1_0 ) fact_sum += factorial(a ) return fact_sum == number if __name__ == "__main__": print("""Program to check whether a number is a Krisnamurthy Number or not.""") SCREAMING_SNAKE_CASE__:Optional[Any] = int(input("""Enter number: """).strip()) print( F'''{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.''' )
67
0
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = pd.read_csv("""sample_data.csv""", header=None) SCREAMING_SNAKE_CASE__:Optional[Any] = df.shape[:1][0] # If you're using some other dataset input the target column SCREAMING_SNAKE_CASE__:Dict = df.iloc[:, 1:2] SCREAMING_SNAKE_CASE__:Optional[int] = actual_data.values.reshape(len_data, 1) SCREAMING_SNAKE_CASE__:Any = MinMaxScaler().fit_transform(actual_data) SCREAMING_SNAKE_CASE__:int = 10 SCREAMING_SNAKE_CASE__:int = 5 SCREAMING_SNAKE_CASE__:List[str] = 20 SCREAMING_SNAKE_CASE__:List[str] = len_data - periods * look_back SCREAMING_SNAKE_CASE__:int = actual_data[:division] SCREAMING_SNAKE_CASE__:List[str] = actual_data[division - look_back :] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__:List[str] = [], [] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__:Tuple = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) SCREAMING_SNAKE_CASE__:int = np.array(train_x) SCREAMING_SNAKE_CASE__:Dict = np.array(test_x) SCREAMING_SNAKE_CASE__:List[Any] = np.array([list(i.ravel()) for i in train_y]) SCREAMING_SNAKE_CASE__:str = np.array([list(i.ravel()) for i in test_y]) SCREAMING_SNAKE_CASE__:Tuple = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") SCREAMING_SNAKE_CASE__:Union[str, Any] = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) SCREAMING_SNAKE_CASE__:Tuple = model.predict(x_test)
703
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) SCREAMING_SNAKE_CASE__:Optional[Any] = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Union[str, Any] = [ """GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTBigCodeForSequenceClassification""", """GPTBigCodeForTokenClassification""", """GPTBigCodeForCausalLM""", """GPTBigCodeModel""", """GPTBigCodePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__:List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = [ ("bert.bert", "visual_bert"), ("bert.cls", "cls"), ("bert.classifier", "cls"), ("token_type_embeddings_visual", "visual_token_type_embeddings"), ("position_embeddings_visual", "visual_position_embeddings"), ("projection", "visual_projection"), ] SCREAMING_SNAKE_CASE__:int = [ "nlvr2_coco_pre_trained.th", "nlvr2_fine_tuned.th", "nlvr2_pre_trained.th", "vcr_coco_pre_train.th", "vcr_fine_tune.th", "vcr_pre_train.th", "vqa_coco_pre_trained.th", "vqa_fine_tuned.th", "vqa_pre_trained.th", ] def _lowerCamelCase( a ): __a = torch.load(a , map_location="cpu" ) return sd def _lowerCamelCase( a , a , a=rename_keys_prefix ): __a = OrderedDict() __a = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue __a = key for name_pair in rename_keys_prefix: __a = new_key.replace(name_pair[0] , name_pair[1] ) __a = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately __a = new_d["cls.predictions.bias"] return new_d @torch.no_grad() def _lowerCamelCase( a , a ): assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), F"The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}." # Get Config if "pre" in checkpoint_path: __a = "pretraining" if "vcr" in checkpoint_path: __a = {"visual_embedding_dim": 5_1_2} elif "vqa_advanced" in checkpoint_path: __a = {"visual_embedding_dim": 2_0_4_8} elif "vqa" in checkpoint_path: __a = {"visual_embedding_dim": 2_0_4_8} elif "nlvr" in checkpoint_path: __a = {"visual_embedding_dim": 1_0_2_4} else: raise NotImplementedError(F"No implementation found for `{checkpoint_path}`." ) else: if "vcr" in checkpoint_path: __a = {"visual_embedding_dim": 5_1_2} __a = "multichoice" elif "vqa_advanced" in checkpoint_path: __a = {"visual_embedding_dim": 2_0_4_8} __a = "vqa_advanced" elif "vqa" in checkpoint_path: __a = {"visual_embedding_dim": 2_0_4_8, "num_labels": 3_1_2_9} __a = "vqa" elif "nlvr" in checkpoint_path: __a = { "visual_embedding_dim": 1_0_2_4, "num_labels": 2, } __a = "nlvr" __a = VisualBertConfig(**a ) # Load State Dict __a = load_state_dict(a ) __a = get_new_dict(a , a ) if model_type == "pretraining": __a = VisualBertForPreTraining(a ) elif model_type == "vqa": __a = VisualBertForQuestionAnswering(a ) elif model_type == "nlvr": __a = VisualBertForVisualReasoning(a ) elif model_type == "multichoice": __a = VisualBertForMultipleChoice(a ) model.load_state_dict(a ) # Save Checkpoints Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Any = argparse.ArgumentParser() # Required parameters parser.add_argument("""orig_checkpoint_path""", type=str, help="""A path to .th on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", type=str, help="""Path to the output PyTorch model.""") SCREAMING_SNAKE_CASE__:Optional[int] = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
704
"""simple docstring""" import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def _lowerCamelCase( a , a , a ): __a = OmegaConf.load(a ) __a = torch.load(a , map_location="cpu" )["model"] __a = list(state_dict.keys() ) # extract state_dict for VQVAE __a = {} __a = "first_stage_model." for key in keys: if key.startswith(a ): __a = state_dict[key] # extract state_dict for UNetLDM __a = {} __a = "model.diffusion_model." for key in keys: if key.startswith(a ): __a = state_dict[key] __a = config.model.params.first_stage_config.params __a = config.model.params.unet_config.params __a = VQModel(**a ).eval() vqvae.load_state_dict(a ) __a = UNetLDMModel(**a ).eval() unet.load_state_dict(a ) __a = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=a , ) __a = LDMPipeline(a , a , a ) pipeline.save_pretrained(a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[Any] = argparse.ArgumentParser() parser.add_argument("""--checkpoint_path""", type=str, required=True) parser.add_argument("""--config_path""", type=str, required=True) parser.add_argument("""--output_path""", type=str, required=True) SCREAMING_SNAKE_CASE__:Union[str, Any] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
67
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class snake_case__ ( __lowercase, unittest.TestCase ): _snake_case : str = ShapEImgaImgPipeline _snake_case : str = ["""image"""] _snake_case : List[str] = ["""image"""] _snake_case : Union[str, Any] = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] _snake_case : List[str] = False @property def a__ ( self ): return 32 @property def a__ ( self ): return 32 @property def a__ ( self ): return self.time_input_dim * 4 @property def a__ ( self ): return 8 @property def a__ ( self ): torch.manual_seed(0 ) __a = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=64 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) __a = CLIPVisionModel(lowerCamelCase ) return model @property def a__ ( self ): __a = CLIPImageProcessor( crop_size=224 , do_center_crop=lowerCamelCase , do_normalize=lowerCamelCase , do_resize=lowerCamelCase , image_mean=[0.4814_5466, 0.457_8275, 0.4082_1073] , image_std=[0.2686_2954, 0.2613_0258, 0.2757_7711] , resample=3 , size=224 , ) return image_processor @property def a__ ( self ): torch.manual_seed(0 ) __a = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "embedding_proj_norm_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } __a = PriorTransformer(**lowerCamelCase ) return model @property def a__ ( self ): torch.manual_seed(0 ) __a = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } __a = ShapERenderer(**lowerCamelCase ) return model def a__ ( self ): __a = self.dummy_prior __a = self.dummy_image_encoder __a = self.dummy_image_processor __a = self.dummy_renderer __a = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1024 , prediction_type="sample" , use_karras_sigmas=lowerCamelCase , clip_sample=lowerCamelCase , clip_sample_range=1.0 , ) __a = { "prior": prior, "image_encoder": image_encoder, "image_processor": image_processor, "renderer": renderer, "scheduler": scheduler, } return components def a__ ( self , lowerCamelCase , lowerCamelCase=0 ): __a = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) if str(lowerCamelCase ).startswith("mps" ): __a = torch.manual_seed(lowerCamelCase ) else: __a = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __a = { "image": input_image, "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def a__ ( self ): __a = "cpu" __a = self.get_dummy_components() __a = self.pipeline_class(**lowerCamelCase ) __a = pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __a = pipe(**self.get_dummy_inputs(lowerCamelCase ) ) __a = output.images[0] __a = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __a = np.array( [ 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, 0.0003_9216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def a__ ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def a__ ( self ): __a = torch_device == "cpu" __a = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=lowerCamelCase , relax_max_difference=lowerCamelCase , ) def a__ ( self ): __a = self.get_dummy_components() __a = self.pipeline_class(**lowerCamelCase ) __a = pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __a = 1 __a = 2 __a = self.get_dummy_inputs(lowerCamelCase ) for key in inputs.keys(): if key in self.batch_params: __a = batch_size * [inputs[key]] __a = pipe(**lowerCamelCase , num_images_per_prompt=lowerCamelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class snake_case__ ( unittest.TestCase ): def a__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def a__ ( self ): __a = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/corgi.png" ) __a = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_img2img_out.npy" ) __a = ShapEImgaImgPipeline.from_pretrained("openai/shap-e-img2img" ) __a = pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __a = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __a = pipe( lowerCamelCase , generator=lowerCamelCase , guidance_scale=3.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(lowerCamelCase , lowerCamelCase )
705
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Optional[Any] = { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/config.json""", # See all BlenderbotSmall models at https://huggingface.co/models?filter=blenderbot_small } class snake_case__ ( snake_case_ ): _snake_case : str = """blenderbot-small""" _snake_case : str = ["""past_key_values"""] _snake_case : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCamelCase=50265 , lowerCamelCase=512 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=8 , lowerCamelCase=2048 , lowerCamelCase=16 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase="gelu" , lowerCamelCase=512 , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=0.0 , lowerCamelCase=0.02 , lowerCamelCase=1 , lowerCamelCase=False , lowerCamelCase=0 , lowerCamelCase=1 , lowerCamelCase=2 , lowerCamelCase=2 , **lowerCamelCase , ): __a = vocab_size __a = max_position_embeddings __a = d_model __a = encoder_ffn_dim __a = encoder_layers __a = encoder_attention_heads __a = decoder_ffn_dim __a = decoder_layers __a = decoder_attention_heads __a = dropout __a = attention_dropout __a = activation_dropout __a = activation_function __a = init_std __a = encoder_layerdrop __a = decoder_layerdrop __a = use_cache __a = encoder_layers __a = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , is_encoder_decoder=lowerCamelCase , decoder_start_token_id=lowerCamelCase , forced_eos_token_id=lowerCamelCase , **lowerCamelCase , ) class snake_case__ ( snake_case_ ): @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a = {0: "batch"} __a = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __a = {0: "batch", 1: "decoder_sequence"} __a = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} else: __a = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def a__ ( self ): if self.task in ["default", "seq2seq-lm"]: __a = super().outputs else: __a = super(lowerCamelCase , self ).outputs if self.use_past: __a , __a = self.num_layers for i in range(lowerCamelCase ): __a = {0: "batch", 2: "past_sequence + sequence"} __a = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # Generate decoder inputs __a = seq_length if not self.use_past else 1 __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) __a = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} __a = dict(**lowerCamelCase , **lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape __a = common_inputs["decoder_input_ids"].shape[1] __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = decoder_seq_length + 3 __a = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __a = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase )] , dim=1 ) __a = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __a , __a = self.num_layers __a = min(lowerCamelCase , lowerCamelCase ) __a = max(lowerCamelCase , lowerCamelCase ) - min_num_layers __a = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(lowerCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase ), ) ) # TODO: test this. __a = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(lowerCamelCase , lowerCamelCase ): common_inputs["past_key_values"].append((torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __a , __a = common_inputs["input_ids"].shape # Not using the same length for past_key_values __a = seqlen + 2 __a , __a = self.num_layers __a , __a = self.num_attention_heads __a = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __a = common_inputs["attention_mask"].dtype __a = torch.cat( [common_inputs["attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) __a = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(lowerCamelCase ) ] return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __a = tokenizer.num_special_tokens_to_add(lowerCamelCase ) __a = compute_effective_axis_dimension( lowerCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCamelCase ) # Generate dummy inputs according to compute batch and sequence __a = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __a = dict(tokenizer(lowerCamelCase , return_tensors=lowerCamelCase ) ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase = -1 , lowerCamelCase = -1 , lowerCamelCase = False , lowerCamelCase = None , ): if self.task in ["default", "seq2seq-lm"]: __a = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) elif self.task == "causal-lm": __a = self._generate_dummy_inputs_for_causal_lm( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) else: __a = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) return common_inputs def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): if self.task in ["default", "seq2seq-lm"]: __a = super()._flatten_past_key_values_(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __a = super(lowerCamelCase , self )._flatten_past_key_values_( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class snake_case__ ( snake_case_ ): def __init__( self , lowerCamelCase=0.01 , lowerCamelCase=1000 ): __a = p_stop __a = max_length def __iter__( self ): __a = 0 __a = False while not stop and count < self.max_length: yield count count += 1 __a = random.random() < self.p_stop class snake_case__ ( unittest.TestCase ): def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=False , lowerCamelCase=True ): __a = [ BatchSamplerShard(__UpperCamelCase , 2 , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) for i in range(2 ) ] __a = [list(__UpperCamelCase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(__UpperCamelCase ) for shard in batch_sampler_shards] , [len(__UpperCamelCase ) for e in expected] ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) def a__ ( self ): # Check the shards when the dataset is a round multiple of total batch size. __a = BatchSampler(range(24 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) __a = BatchSampler(range(24 ) , batch_size=3 , drop_last=__UpperCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. __a = BatchSampler(range(21 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) __a = BatchSampler(range(21 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. __a = BatchSampler(range(22 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) __a = BatchSampler(range(22 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. __a = BatchSampler(range(20 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) __a = BatchSampler(range(20 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) # Check the shards when the dataset is very small. __a = BatchSampler(range(2 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) __a = BatchSampler(range(2 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [[], []] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase ) def a__ ( self ): # Check the shards when the dataset is a round multiple of batch size. __a = BatchSampler(range(24 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) __a = BatchSampler(range(24 ) , batch_size=4 , drop_last=__UpperCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. __a = BatchSampler(range(22 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) __a = BatchSampler(range(22 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. __a = BatchSampler(range(21 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) __a = BatchSampler(range(21 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) # Check the shards when the dataset is very small. __a = BatchSampler(range(2 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) __a = BatchSampler(range(2 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [[], []] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase ) def a__ ( self ): # Check the shards when the dataset is a round multiple of total batch size. __a = BatchSampler(range(24 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(24 ) , batch_size=3 , drop_last=__UpperCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. __a = BatchSampler(range(21 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(21 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. __a = BatchSampler(range(22 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(22 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. __a = BatchSampler(range(20 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(20 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is very small. __a = BatchSampler(range(2 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [[[0, 1]], []] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(2 ) , batch_size=3 , drop_last=__UpperCamelCase ) __a = [[], []] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , even_batches=__UpperCamelCase ) def a__ ( self ): # Check the shards when the dataset is a round multiple of batch size. __a = BatchSampler(range(24 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(24 ) , batch_size=4 , drop_last=__UpperCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. __a = BatchSampler(range(22 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(22 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. __a = BatchSampler(range(21 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(21 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) # Check the shards when the dataset is very small. __a = BatchSampler(range(2 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [[[0, 1]], []] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) __a = BatchSampler(range(2 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = [[], []] self.check_batch_sampler_shards(__UpperCamelCase , __UpperCamelCase , split_batches=__UpperCamelCase , even_batches=__UpperCamelCase ) def a__ ( self ): __a = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] __a = [BatchSamplerShard(__UpperCamelCase , 2 , __UpperCamelCase , even_batches=__UpperCamelCase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=False , lowerCamelCase=2 , lowerCamelCase=False ): random.seed(__UpperCamelCase ) __a = list(__UpperCamelCase ) __a = [ IterableDatasetShard( __UpperCamelCase , batch_size=__UpperCamelCase , drop_last=__UpperCamelCase , num_processes=__UpperCamelCase , process_index=__UpperCamelCase , split_batches=__UpperCamelCase , ) for i in range(__UpperCamelCase ) ] __a = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(__UpperCamelCase ) iterable_dataset_lists.append(list(__UpperCamelCase ) ) __a = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size __a = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(__UpperCamelCase ) , len(__UpperCamelCase ) ) self.assertTrue(len(__UpperCamelCase ) % shard_batch_size == 0 ) __a = [] for idx in range(0 , len(__UpperCamelCase ) , __UpperCamelCase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(__UpperCamelCase ) < len(__UpperCamelCase ): reference += reference self.assertListEqual(__UpperCamelCase , reference[: len(__UpperCamelCase )] ) def a__ ( self ): __a = 42 __a = RandomIterableDataset() self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) # Edge case with a very small dataset __a = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) self.check_iterable_dataset_shards(__UpperCamelCase , __UpperCamelCase , batch_size=4 , drop_last=__UpperCamelCase , split_batches=__UpperCamelCase ) def a__ ( self ): __a = BatchSampler(range(16 ) , batch_size=4 , drop_last=__UpperCamelCase ) __a = SkipBatchSampler(__UpperCamelCase , 2 ) self.assertListEqual(list(__UpperCamelCase ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def a__ ( self ): __a = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def a__ ( self ): __a = DataLoader(list(range(16 ) ) , batch_size=4 ) __a = skip_first_batches(__UpperCamelCase , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def a__ ( self ): __a = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(__UpperCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__UpperCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def a__ ( self ): Accelerator() __a = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(__UpperCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__UpperCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
706
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=99 , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=9 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase=8 , lowerCamelCase=0.1 , lowerCamelCase=0.002 , lowerCamelCase=1 , lowerCamelCase=0 , lowerCamelCase=0 , lowerCamelCase=None , lowerCamelCase=None , ): __a = parent __a = batch_size __a = encoder_seq_length __a = decoder_seq_length # For common tests __a = self.decoder_seq_length __a = is_training __a = use_attention_mask __a = use_labels __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = d_ff __a = relative_attention_num_buckets __a = dropout_rate __a = initializer_factor __a = eos_token_id __a = pad_token_id __a = decoder_start_token_id __a = None __a = decoder_layers def a__ ( self ): return TaConfig.from_pretrained("google/umt5-base" ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , ): if attention_mask is None: __a = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __a = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __a = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=lowerCamelCase ) if decoder_head_mask is None: __a = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) if cross_attn_head_mask is None: __a = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=lowerCamelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def a__ ( self ): __a = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) __a = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __a = input_ids.clamp(self.pad_token_id + 1 ) __a = decoder_input_ids.clamp(self.pad_token_id + 1 ) __a = self.get_config() __a = config.num_attention_heads __a = self.prepare_inputs_dict(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return config, input_dict def a__ ( self ): __a , __a = self.prepare_config_and_inputs() return config, inputs_dict def a__ ( self ): return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self ): return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase , attention_mask=lowerCamelCase , decoder_attention_mask=lowerCamelCase , ) __a = model(input_ids=lowerCamelCase , decoder_input_ids=lowerCamelCase ) __a = result.last_hidden_state __a = result.past_key_values __a = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(lowerCamelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).get_decoder().to(lowerCamelCase ).eval() # first forward pass __a = model(lowerCamelCase , use_cache=lowerCamelCase ) __a = model(lowerCamelCase ) __a = model(lowerCamelCase , use_cache=lowerCamelCase ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) ) self.parent.assertTrue(len(lowerCamelCase ) == len(lowerCamelCase ) + 1 ) __a , __a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = model(lowerCamelCase )["last_hidden_state"] __a = model(lowerCamelCase , past_key_values=lowerCamelCase )["last_hidden_state"] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -1, random_slice_idx].detach() __a = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , ): __a = UMTaModel(config=lowerCamelCase ).to(lowerCamelCase ).half().eval() __a = model(**lowerCamelCase )["last_hidden_state"] self.parent.assertFalse(torch.isnan(lowerCamelCase ).any().item() ) @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _snake_case : int = (UMTaForConditionalGeneration,) if is_torch_available() else () _snake_case : Optional[int] = ( { """conversational""": UMTaForConditionalGeneration, """feature-extraction""": UMTaModel, """summarization""": UMTaForConditionalGeneration, """text2text-generation""": UMTaForConditionalGeneration, """translation""": UMTaForConditionalGeneration, """question-answering""": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _snake_case : List[Any] = True _snake_case : Union[str, Any] = False _snake_case : Union[str, Any] = False _snake_case : Tuple = True _snake_case : List[str] = True # The small UMT5 model needs higher percentages for CPU/MP tests _snake_case : Optional[Any] = [0.8, 0.9] def a__ ( self ): __a = UMTaModelTester(self ) @unittest.skip("Test has a segmentation fault on torch 1.8.0" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() __a = UMTaModel(config_and_inputs[0] ).to(lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( lowerCamelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , F"{tmpdirname}/t5_test.onnx" , export_params=lowerCamelCase , opset_version=9 , input_names=["input_ids", "decoder_input_ids"] , ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*lowerCamelCase ) def a__ ( self ): __a = ["encoder_attentions", "decoder_attentions", "cross_attentions"] __a = self.model_tester.prepare_config_and_inputs() __a = config_and_inputs[0] __a = UMTaForConditionalGeneration(lowerCamelCase ).eval() model.to(lowerCamelCase ) __a = { "head_mask": torch.zeros(config.num_layers , config.num_heads , device=lowerCamelCase ), "decoder_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), "cross_attn_head_mask": torch.zeros(config.num_decoder_layers , config.num_heads , device=lowerCamelCase ), } for attn_name, (name, mask) in zip(lowerCamelCase , head_masking.items() ): __a = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": __a = torch.ones( config.num_decoder_layers , config.num_heads , device=lowerCamelCase ) __a = model.generate( config_and_inputs[1]["input_ids"] , num_beams=1 , max_length=3 , output_attentions=lowerCamelCase , return_dict_in_generate=lowerCamelCase , **lowerCamelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step __a = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("Does not work on the tiny model as we keep hitting edge cases." ) def a__ ( self ): pass @require_torch @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow @unittest.skip( "Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged" ) def a__ ( self ): __a = UMTaForConditionalGeneration.from_pretrained("google/umt5-small" , return_dict=lowerCamelCase ).to(lowerCamelCase ) __a = AutoTokenizer.from_pretrained("google/umt5-small" , use_fast=lowerCamelCase , legacy=lowerCamelCase ) __a = [ "Bonjour monsieur <extra_id_0> bien <extra_id_1>.", "No se como puedo <extra_id_0>.", "This is the reason why we <extra_id_0> them.", "The <extra_id_0> walks in <extra_id_1>, seats", "A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.", ] __a = tokenizer(lowerCamelCase , return_tensors="pt" , padding=lowerCamelCase ).input_ids # fmt: off __a = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(lowerCamelCase , lowerCamelCase ) __a = model.generate(input_ids.to(lowerCamelCase ) ) __a = [ "<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>", "<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", "<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>", ] __a = tokenizer.batch_decode(lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase )
67
0
"""simple docstring""" import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin __lowerCamelCase:Optional[int] = get_tests_dir("""fixtures/spiece.model""") @require_sentencepiece @require_tokenizers class snake_case__ ( A_, unittest.TestCase ): _snake_case : int = AlbertTokenizer _snake_case : Optional[int] = AlbertTokenizerFast _snake_case : Tuple = True _snake_case : Any = True _snake_case : Optional[int] = True def a__ ( self ): super().setUp() # We have a SentencePiece fixture for testing __a = AlbertTokenizer(lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def a__ ( self , lowerCamelCase ): __a = """this is a test""" __a = """this is a test""" return input_text, output_text def a__ ( self ): __a = """<pad>""" __a = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase ) , lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase ) , lowerCamelCase ) def a__ ( self ): __a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<pad>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "▁eloquent" ) self.assertEqual(len(lowerCamelCase ) , 30000 ) def a__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 30000 ) def a__ ( self ): if not self.test_rust_tokenizer: return __a = self.get_tokenizer() __a = self.get_rust_tokenizer() __a = """I was born in 92000, and this is falsé.""" __a = tokenizer.tokenize(lowerCamelCase ) __a = rust_tokenizer.tokenize(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __a = tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) __a = rust_tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) __a = self.get_rust_tokenizer() __a = tokenizer.encode(lowerCamelCase ) __a = rust_tokenizer.encode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = AlbertTokenizer(lowerCamelCase , keep_accents=lowerCamelCase ) __a = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowerCamelCase , ["▁this", "▁is", "▁a", "▁test"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase ) , [48, 25, 21, 1289] ) __a = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowerCamelCase , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "."] ) __a = tokenizer.convert_tokens_to_ids(lowerCamelCase ) self.assertListEqual(lowerCamelCase , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) __a = tokenizer.convert_ids_to_tokens(lowerCamelCase ) self.assertListEqual( lowerCamelCase , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "."] , ) def a__ ( self ): __a = AlbertTokenizer(lowerCamelCase ) __a = tokenizer.encode("sequence builders" ) __a = tokenizer.encode("multi-sequence build" ) __a = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) __a = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def a__ ( self ): # fmt: off __a = {"""attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """input_ids""": [[2, 21970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 12051, 18, 17, 7103, 2153, 673, 8, 3515, 18684, 8, 4461, 6, 1927, 297, 8, 12060, 2607, 18, 13, 5, 4461, 15, 10538, 38, 8, 135, 15, 822, 58, 15, 993, 10363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 10641, 6, 29, 84, 2512, 2430, 782, 18684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 11712, 15, 7103, 2153, 673, 17, 24883, 9990, 9, 3], [2, 11502, 25, 1006, 20, 782, 8, 11809, 855, 1732, 19393, 18667, 37, 367, 21018, 69, 1854, 34, 11860, 19124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 17659, 84, 14, 16792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """token_type_ids""": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase , model_name="albert-base-v2" , revision="6b6560eaf5ff2e250b00c50f380c5389a9c2d82e" , )
707
"""simple docstring""" import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase( a , a , a ): # Initialise PyTorch model __a = MobileBertConfig.from_json_file(a ) print(F"Building PyTorch model from configuration: {config}" ) __a = MobileBertForPreTraining(a ) # Load weights from tf checkpoint __a = load_tf_weights_in_mobilebert(a , a , a ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , a ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) SCREAMING_SNAKE_CASE__:List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
67
0
import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def _lowerCamelCase( a , a ): assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def _lowerCamelCase( a , a , a ): __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __a = JsonDatasetReader(lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , keep_in_memory=lowerCAmelCase__ ).read() _check_json_dataset(lowerCAmelCase__ , lowerCAmelCase__ ) @pytest.mark.parametrize( "features" , [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ] , ) def _lowerCamelCase( a , a , a ): __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} __a = features.copy() if features else default_expected_features __a = ( Features({feature: Value(lowerCAmelCase__ ) for feature, dtype in features.items()} ) if features is not None else None ) __a = JsonDatasetReader(lowerCAmelCase__ , features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ ).read() _check_json_dataset(lowerCAmelCase__ , lowerCAmelCase__ ) @pytest.mark.parametrize( "features" , [ None, {"col_3": "float64", "col_1": "string", "col_2": "int64"}, ] , ) def _lowerCamelCase( a , a , a ): __a = tmp_path / 'cache' __a = {'col_3': 'float64', 'col_1': 'string', 'col_2': 'int64'} __a = features.copy() if features else default_expected_features __a = ( Features({feature: Value(lowerCAmelCase__ ) for feature, dtype in features.items()} ) if features is not None else None ) __a = JsonDatasetReader(lowerCAmelCase__ , features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ ).read() assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def _lowerCamelCase( a , a ): __a = {'col_2': 'int64', 'col_3': 'float64', 'col_1': 'string'} __a = features.copy() __a = ( Features({feature: Value(lowerCAmelCase__ ) for feature, dtype in features.items()} ) if features is not None else None ) __a = tmp_path / 'cache' __a = JsonDatasetReader(lowerCAmelCase__ , features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ ).read() assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def _lowerCamelCase( a , a , a ): __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} __a = JsonDatasetReader(lowerCAmelCase__ , cache_dir=lowerCAmelCase__ , split=lowerCAmelCase__ ).read() _check_json_dataset(lowerCAmelCase__ , lowerCAmelCase__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list] ) def _lowerCamelCase( a , a , a ): if issubclass(lowerCAmelCase__ , lowerCAmelCase__ ): __a = jsonl_path elif issubclass(lowerCAmelCase__ , lowerCAmelCase__ ): __a = [jsonl_path] __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} __a = JsonDatasetReader(lowerCAmelCase__ , cache_dir=lowerCAmelCase__ ).read() _check_json_dataset(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowerCamelCase( a , a , a=("train",) ): assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) for split in splits: __a = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True] ) def _lowerCamelCase( a , a , a ): __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): __a = JsonDatasetReader({"train": jsonl_path} , cache_dir=lowerCAmelCase__ , keep_in_memory=lowerCAmelCase__ ).read() _check_json_datasetdict(lowerCAmelCase__ , lowerCAmelCase__ ) @pytest.mark.parametrize( "features" , [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ] , ) def _lowerCamelCase( a , a , a ): __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} __a = features.copy() if features else default_expected_features __a = ( Features({feature: Value(lowerCAmelCase__ ) for feature, dtype in features.items()} ) if features is not None else None ) __a = JsonDatasetReader({"train": jsonl_path} , features=lowerCAmelCase__ , cache_dir=lowerCAmelCase__ ).read() _check_json_datasetdict(lowerCAmelCase__ , lowerCAmelCase__ ) @pytest.mark.parametrize("split" , [None, NamedSplit("train" ), "train", "test"] ) def _lowerCamelCase( a , a , a ): if split: __a = {split: jsonl_path} else: __a = 'train' __a = {'train': jsonl_path, 'test': jsonl_path} __a = tmp_path / 'cache' __a = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} __a = JsonDatasetReader(lowerCAmelCase__ , cache_dir=lowerCAmelCase__ ).read() _check_json_datasetdict(lowerCAmelCase__ , lowerCAmelCase__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def _lowerCamelCase( a ): return json.load(lowerCAmelCase__ ) def _lowerCamelCase( a ): return [json.loads(lowerCAmelCase__ ) for line in buffer] class snake_case__ : @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase__ , UpperCamelCase__ , lines=UpperCamelCase__ ).write() buffer.seek(0 ) __a = load_json_function(UpperCamelCase__ ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) assert isinstance(exported_content[0] , UpperCamelCase__ ) assert len(UpperCamelCase__ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase__ , UpperCamelCase__ , lines=UpperCamelCase__ , orient=UpperCamelCase__ ).write() buffer.seek(0 ) __a = load_json(UpperCamelCase__ ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(UpperCamelCase__ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(UpperCamelCase__ ) == 10 @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase__ , UpperCamelCase__ , lines=UpperCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) __a = load_json_function(UpperCamelCase__ ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) assert isinstance(exported_content[0] , UpperCamelCase__ ) assert len(UpperCamelCase__ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase__ , UpperCamelCase__ , lines=UpperCamelCase__ , orient=UpperCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) __a = load_json(UpperCamelCase__ ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(UpperCamelCase__ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(UpperCamelCase__ ) == 10 def a__ ( self , lowerCamelCase ): with pytest.raises(UpperCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(UpperCamelCase__ , UpperCamelCase__ , num_proc=0 ) @pytest.mark.parametrize("compression, extension" , [("gzip", "gz"), ("bz2", "bz2"), ("xz", "xz")] ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = tmp_path_factory.mktemp("data" ) / F"test.json.{extension}" __a = str(shared_datadir / F"test_file.json.{extension}" ) JsonDatasetWriter(UpperCamelCase__ , UpperCamelCase__ , compression=UpperCamelCase__ ).write() with fsspec.open(UpperCamelCase__ , "rb" , compression="infer" ) as f: __a = f.read() with fsspec.open(UpperCamelCase__ , "rb" , compression="infer" ) as f: __a = f.read() assert exported_content == original_content
708
"""simple docstring""" import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class snake_case__ ( snake_case_ ): def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __a = input_file.read() __a = regexp.search(lowerCamelCase ) return match def a__ ( self , lowerCamelCase ): with open(lowerCamelCase , encoding="utf-8" ) as input_file: __a = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __a = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __a = regexp.finditer(lowerCamelCase ) __a = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCamelCase ) ): raise AssertionError(F"open(...) must use utf-8 encoding in {dataset}" ) def a__ ( self ): __a = Path("./datasets" ) __a = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCamelCase ) ): raise AssertionError(F"print statement found in {dataset}. Use datasets.logger/logging instead." )
67
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE__:List[str] = { """configuration_distilbert""": [ """DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DistilBertConfig""", """DistilBertOnnxConfig""", ], """tokenization_distilbert""": ["""DistilBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Optional[Any] = ["""DistilBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:int = [ """DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """DistilBertForMaskedLM""", """DistilBertForMultipleChoice""", """DistilBertForQuestionAnswering""", """DistilBertForSequenceClassification""", """DistilBertForTokenClassification""", """DistilBertModel""", """DistilBertPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Tuple = [ """TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDistilBertForMaskedLM""", """TFDistilBertForMultipleChoice""", """TFDistilBertForQuestionAnswering""", """TFDistilBertForSequenceClassification""", """TFDistilBertForTokenClassification""", """TFDistilBertMainLayer""", """TFDistilBertModel""", """TFDistilBertPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__:Any = [ """FlaxDistilBertForMaskedLM""", """FlaxDistilBertForMultipleChoice""", """FlaxDistilBertForQuestionAnswering""", """FlaxDistilBertForSequenceClassification""", """FlaxDistilBertForTokenClassification""", """FlaxDistilBertModel""", """FlaxDistilBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__:Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
709
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("""To use the rich extension, install rich with `pip install rich`""")
67
0
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings SCREAMING_SNAKE_CASE__:Union[str, Any] = R""" [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `\" / \"`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `\" // \"`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `\"wiki_dpr\"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `\"train\"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `\"compressed\"`) The index name of the index associated with the `dataset`. One can choose between `\"legacy\"`, `\"exact\"` and `\"compressed\"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a \"dummy\" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. """ @add_start_docstrings(lowercase_ ) class snake_case__ ( lowercase_ ): _snake_case : List[str] = '''rag''' _snake_case : Tuple = True def __init__( self , lowerCamelCase=None , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=" / " , lowerCamelCase=" // " , lowerCamelCase=5 , lowerCamelCase=300 , lowerCamelCase=768 , lowerCamelCase=8 , lowerCamelCase="wiki_dpr" , lowerCamelCase="train" , lowerCamelCase="compressed" , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=True , lowerCamelCase=None , **lowerCamelCase , ): super().__init__( bos_token_id=UpperCamelCase__ , pad_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , decoder_start_token_id=UpperCamelCase__ , forced_eos_token_id=UpperCamelCase__ , is_encoder_decoder=UpperCamelCase__ , prefix=UpperCamelCase__ , vocab_size=UpperCamelCase__ , **UpperCamelCase__ , ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" __a = kwargs.pop("question_encoder" ) __a = question_encoder_config.pop("model_type" ) __a = kwargs.pop("generator" ) __a = decoder_config.pop("model_type" ) from ..auto.configuration_auto import AutoConfig __a = AutoConfig.for_model(UpperCamelCase__ , **UpperCamelCase__ ) __a = AutoConfig.for_model(UpperCamelCase__ , **UpperCamelCase__ ) __a = reduce_loss __a = label_smoothing __a = exclude_bos_score __a = do_marginalize __a = title_sep __a = doc_sep __a = n_docs __a = max_combined_length __a = dataset __a = dataset_split __a = index_name __a = retrieval_vector_size __a = retrieval_batch_size __a = passages_path __a = index_path __a = use_dummy_dataset __a = output_retrieved __a = do_deduplication __a = use_cache if self.forced_eos_token_id is None: __a = getattr(self.generator , "forced_eos_token_id" , UpperCamelCase__ ) @classmethod def a__ ( cls , lowerCamelCase , lowerCamelCase , **lowerCamelCase ): return cls(question_encoder=question_encoder_config.to_dict() , generator=generator_config.to_dict() , **UpperCamelCase__ ) def a__ ( self ): __a = copy.deepcopy(self.__dict__ ) __a = self.question_encoder.to_dict() __a = self.generator.to_dict() __a = self.__class__.model_type return output
710
"""simple docstring""" import heapq import sys import numpy as np SCREAMING_SNAKE_CASE__:Optional[int] = tuple[int, int] class snake_case__ : def __init__( self ): __a = [] __a = set() def a__ ( self ): if not self.empty(): return self.elements[0][0] else: return float("inf" ) def a__ ( self ): return len(self.elements ) == 0 def a__ ( self , lowerCamelCase , lowerCamelCase ): if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(lowerCamelCase ) else: # update # print("update", item) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def a__ ( self , lowerCamelCase ): if item in self.set: self.set.remove(lowerCamelCase ) __a = [] ((__a) , (__a)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__a) , (__a)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def a__ ( self ): return self.elements[0][1] def a__ ( self ): ((__a) , (__a)) = heapq.heappop(self.elements ) self.set.remove(lowerCamelCase ) return (priority, item) def _lowerCamelCase( a , a ): # euclidean distance __a = np.array(a ) __a = np.array(a ) return np.linalg.norm(a - b ) def _lowerCamelCase( a , a ): # integer division by time variable return consistent_heuristic(a , a ) // t def _lowerCamelCase( a , a ): # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _lowerCamelCase( a , a , a , a ): __a = g_function[start] + Wa * heuristics[i](a , a ) return ans def _lowerCamelCase( a , a , a ): __a = np.chararray((n, n) ) for i in range(a ): for j in range(a ): __a = "*" for i in range(a ): for j in range(a ): if (j, (n - 1) - i) in blocks: __a = "#" __a = "-" __a = back_pointer[goal] while x != start: ((__a) , (__a)) = x # print(x) __a = "-" __a = back_pointer[x] __a = "-" for i in range(a ): for j in range(a ): if (i, j) == (0, n - 1): print(grid[i][j] , end=" " ) print("<-- End position" , end=" " ) else: print(grid[i][j] , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) print("PATH TAKEN BY THE ALGORITHM IS:-" ) __a = back_pointer[goal] while x != start: print(a , end=" " ) __a = back_pointer[x] print(a ) sys.exit() def _lowerCamelCase( a ): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _lowerCamelCase( a , a , a , a , a , a , a , a , ): for itera in range(a ): open_list[itera].remove_element(a ) # print("s", s) # print("j", j) ((__a) , (__a)) = s __a = (x - 1, y) __a = (x + 1, y) __a = (x, y + 1) __a = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(a ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(a ) __a = -1 __a = float("inf" ) if valid(a ) and g_function[neighbours] > g_function[s] + 1: __a = g_function[s] + 1 __a = s if neighbours not in close_list_anchor: open_list[0].put(a , key(a , 0 , a , a ) ) if neighbours not in close_list_inad: for var in range(1 , a ): if key(a , a , a , a ) <= Wa * key( a , 0 , a , a ): open_list[j].put( a , key(a , a , a , a ) ) def _lowerCamelCase( ): __a = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(1_5 , 2_0 ): some_list.append((x, 1_7) ) for x in range(1_0 , 1_9 ): for y in range(1 , 1_5 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(1_2 , 1_9 ): some_list.append((x, y) ) for x in range(3 , 1_3 ): for y in range(1_6 , 1_9 ): some_list.append((x, y) ) return some_list SCREAMING_SNAKE_CASE__:Any = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} SCREAMING_SNAKE_CASE__:str = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] SCREAMING_SNAKE_CASE__:int = make_common_ground() SCREAMING_SNAKE_CASE__:List[str] = blocks_blk # hyper parameters SCREAMING_SNAKE_CASE__:str = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 1 SCREAMING_SNAKE_CASE__:Union[str, Any] = 20 SCREAMING_SNAKE_CASE__:Dict = 3 # one consistent and two other inconsistent # start and end destination SCREAMING_SNAKE_CASE__:Dict = (0, 0) SCREAMING_SNAKE_CASE__:Optional[Any] = (n - 1, n - 1) SCREAMING_SNAKE_CASE__:List[str] = 1 def _lowerCamelCase( a , a , a ): __a = {start: 0, goal: float("inf" )} __a = {start: -1, goal: -1} __a = [] __a = set() for i in range(a ): open_list.append(PriorityQueue() ) open_list[i].put(a , key(a , a , a , a ) ) __a = [] __a = [] while open_list[0].minkey() < float("inf" ): for i in range(1 , a ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a , __a = open_list[i].top_show() visited.add(a ) expand_state( a , a , a , a , a , a , a , a , ) close_list_inad.append(a ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf" ): do_something(a , a , a ) else: __a = open_list[0].top_show() visited.add(a ) expand_state( a , 0 , a , a , a , a , a , a , ) close_list_anchor.append(a ) print("No path found to goal" ) print() for i in range(n - 1 , -1 , -1 ): for j in range(a ): if (j, i) in blocks: print("#" , end=" " ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" " ) else: print("-" , end=" " ) else: print("*" , end=" " ) if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
67
0
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class snake_case__ ( __UpperCAmelCase ): _snake_case : int = 42 class snake_case__ ( __UpperCAmelCase, __UpperCAmelCase ): @register_to_config def __init__( self , lowerCamelCase = 65536 , lowerCamelCase = None , lowerCamelCase = 2 , lowerCamelCase = 2 , lowerCamelCase = 0 , lowerCamelCase = "fourier" , lowerCamelCase = True , lowerCamelCase = False , lowerCamelCase = 0.0 , lowerCamelCase = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , lowerCamelCase = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , lowerCamelCase = "UNetMidBlock1D" , lowerCamelCase = None , lowerCamelCase = (32, 32, 64) , lowerCamelCase = None , lowerCamelCase = 8 , lowerCamelCase = 1 , lowerCamelCase = False , ): super().__init__() __a = sample_size # time if time_embedding_type == "fourier": __a = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=lowerCAmelCase_ , log=lowerCAmelCase_ , flip_sin_to_cos=lowerCAmelCase_ ) __a = 2 * block_out_channels[0] elif time_embedding_type == "positional": __a = Timesteps( block_out_channels[0] , flip_sin_to_cos=lowerCAmelCase_ , downscale_freq_shift=lowerCAmelCase_ ) __a = block_out_channels[0] if use_timestep_embedding: __a = block_out_channels[0] * 4 __a = TimestepEmbedding( in_channels=lowerCAmelCase_ , time_embed_dim=lowerCAmelCase_ , act_fn=lowerCAmelCase_ , out_dim=block_out_channels[0] , ) __a = nn.ModuleList([] ) __a = None __a = nn.ModuleList([] ) __a = None # down __a = in_channels for i, down_block_type in enumerate(lowerCAmelCase_ ): __a = output_channel __a = block_out_channels[i] if i == 0: input_channel += extra_in_channels __a = i == len(lowerCAmelCase_ ) - 1 __a = get_down_block( lowerCAmelCase_ , num_layers=lowerCAmelCase_ , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(lowerCAmelCase_ ) # mid __a = get_mid_block( lowerCAmelCase_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=lowerCAmelCase_ , add_downsample=lowerCAmelCase_ , ) # up __a = list(reversed(lowerCAmelCase_ ) ) __a = reversed_block_out_channels[0] if out_block_type is None: __a = out_channels else: __a = block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): __a = output_channel __a = ( reversed_block_out_channels[i + 1] if i < len(lowerCAmelCase_ ) - 1 else final_upsample_channels ) __a = i == len(lowerCAmelCase_ ) - 1 __a = get_up_block( lowerCAmelCase_ , num_layers=lowerCAmelCase_ , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(lowerCAmelCase_ ) __a = output_channel # out __a = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) __a = get_out_block( out_block_type=lowerCAmelCase_ , num_groups_out=lowerCAmelCase_ , embed_dim=block_out_channels[0] , out_channels=lowerCAmelCase_ , act_fn=lowerCAmelCase_ , fc_dim=block_out_channels[-1] // 4 , ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = True , ): __a = timestep if not torch.is_tensor(lowerCAmelCase_ ): __a = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(lowerCAmelCase_ ) and len(timesteps.shape ) == 0: __a = timesteps[None].to(sample.device ) __a = self.time_proj(lowerCAmelCase_ ) if self.config.use_timestep_embedding: __a = self.time_mlp(lowerCAmelCase_ ) else: __a = timestep_embed[..., None] __a = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) __a = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down __a = () for downsample_block in self.down_blocks: __a , __a = downsample_block(hidden_states=lowerCAmelCase_ , temb=lowerCAmelCase_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: __a = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): __a = down_block_res_samples[-1:] __a = down_block_res_samples[:-1] __a = upsample_block(lowerCAmelCase_ , res_hidden_states_tuple=lowerCAmelCase_ , temb=lowerCAmelCase_ ) # 5. post-process if self.out_block: __a = self.out_block(lowerCAmelCase_ , lowerCAmelCase_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=lowerCAmelCase_ )
711
"""simple docstring""" SCREAMING_SNAKE_CASE__:Any = """Alexander Joslin""" import operator as op from .stack import Stack def _lowerCamelCase( a ): __a = {"*": op.mul, "/": op.truediv, "+": op.add, "-": op.sub} __a = Stack() __a = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(a ) ) elif i in operators: # RULE 2 operator_stack.push(a ) elif i == ")": # RULE 4 __a = operator_stack.peek() operator_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operand_stack.peek() operand_stack.pop() __a = operators[opr](a , a ) operand_stack.push(a ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
67
0
"""simple docstring""" from collections import deque class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = process_name # process name __a = arrival_time # arrival time of the process # completion time of finished process or last interrupted time __a = arrival_time __a = burst_time # remaining burst time __a = 0 # total time of the process wait in ready queue __a = 0 # time from arrival time to completion time class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ): # total number of mlfq's queues __a = number_of_queues # time slice of queues that round robin algorithm applied __a = time_slices # unfinished process is in this ready_queue __a = queue # current time __a = current_time # finished process is in this sequence queue __a = deque() def a__ ( self ): __a = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def a__ ( self , lowerCamelCase ): __a = [] for i in range(len(_UpperCAmelCase ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def a__ ( self , lowerCamelCase ): __a = [] for i in range(len(_UpperCAmelCase ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def a__ ( self , lowerCamelCase ): __a = [] for i in range(len(_UpperCAmelCase ) ): completion_times.append(queue[i].stop_time ) return completion_times def a__ ( self , lowerCamelCase ): return [q.burst_time for q in queue] def a__ ( self , lowerCamelCase ): process.waiting_time += self.current_time - process.stop_time return process.waiting_time def a__ ( self , lowerCamelCase ): __a = deque() # sequence deque of finished process while len(_UpperCAmelCase ) != 0: __a = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(_UpperCAmelCase ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 __a = 0 # set the process's turnaround time because it is finished __a = self.current_time - cp.arrival_time # set the completion time __a = self.current_time # add the process to queue that has finished queue finished.append(_UpperCAmelCase ) self.finish_queue.extend(_UpperCAmelCase ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def a__ ( self , lowerCamelCase , lowerCamelCase ): __a = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(_UpperCAmelCase ) ): __a = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(_UpperCAmelCase ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time __a = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(_UpperCAmelCase ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished __a = 0 # set the finish time __a = self.current_time # update the process' turnaround time because it is finished __a = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(_UpperCAmelCase ) self.finish_queue.extend(_UpperCAmelCase ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def a__ ( self ): # all queues except last one have round_robin algorithm for i in range(self.number_of_queues - 1 ): __a = self.round_robin( self.ready_queue , self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest SCREAMING_SNAKE_CASE__:str = Process("""P1""", 0, 53) SCREAMING_SNAKE_CASE__:str = Process("""P2""", 0, 17) SCREAMING_SNAKE_CASE__:str = Process("""P3""", 0, 68) SCREAMING_SNAKE_CASE__:Optional[int] = Process("""P4""", 0, 24) SCREAMING_SNAKE_CASE__:Union[str, Any] = 3 SCREAMING_SNAKE_CASE__:Tuple = [17, 25] SCREAMING_SNAKE_CASE__:int = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={"""queue""": deque([Pa, Pa, Pa, Pa])}) SCREAMING_SNAKE_CASE__:List[str] = Process("""P1""", 0, 53) SCREAMING_SNAKE_CASE__:Union[str, Any] = Process("""P2""", 0, 17) SCREAMING_SNAKE_CASE__:List[str] = Process("""P3""", 0, 68) SCREAMING_SNAKE_CASE__:Tuple = Process("""P4""", 0, 24) SCREAMING_SNAKE_CASE__:Union[str, Any] = 3 SCREAMING_SNAKE_CASE__:Optional[Any] = [17, 25] SCREAMING_SNAKE_CASE__:Tuple = deque([Pa, Pa, Pa, Pa]) SCREAMING_SNAKE_CASE__:Union[str, Any] = MLFQ(number_of_queues, time_slices, queue, 0) SCREAMING_SNAKE_CASE__:List[Any] = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F'''waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print completion times of processes(P1, P2, P3, P4) print( F'''completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print total turnaround times of processes(P1, P2, P3, P4) print( F'''turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}''' ) # print sequence of finished processes print( F'''sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}''' )
712
"""simple docstring""" from math import pi def _lowerCamelCase( a , a ): return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
67
0